Glossary of Key Terms
| Artificial Intelligence (AI) | A computer system that can simulate or perform human tasks. | 
| Machine Learning (ML) | A discipline within artificial intelligence (AI) dedicated to the development of algorithms that improve their performance on a human task without requiring explicit instructions. | 
| ML Algorithm/Model | A learned representation of the patterns inherent with the input data that can be used to generate predictions. | 
| Pipeline | A full sequence of steps to convert input data into output predictions. Typically involving loading, reformatting, and transforming data and predictions so that they can be integrating into real-time workflows. | 
| Development Operations (DevOps) | A philosophical framework combining best practices in information technology operations and software engineering to rapidly and robustly build and implement high-quality informatics solutions. | 
| Machine Learning Operations (MLOps) | A framework of technical best practices for deploying and maintaining machine learning applications efficiently and effectively. | 
| Target/Ground-Truth | The “gold-standard” definition to which ML pipeline predictions will be compared. | 
| Discriminative Performance | The degree to which predictions match the ground-truth labels, often measured by sensitivity, specificity, positive predictive value, and negative predictive value. | 
| Implementation Efficacy | The degree to which the final implementation of the ML pipeline satisfies the original need for which it was built. | 
| Receiver Operating Characteristic (ROC Curve) | A visualization of the trade-off between sensitivity and specificity across the full breadth of possible decision thresholds for a continuous output. | 
| Class Imbalance | The degree to which the proportion of class labels are skewed towards one label or the other. | 
| Precision and Recall Curve | A visualization of the trade-off between precision (positive predictive value) and recall (sensitivity) across the full breadth of possible decision thresholds for a continuous output. | 
| F1 Score | The harmonic mean of precision and recall. See more here. | 
| Matthews Correlation Coefficient | The Pearson correlation coefficient for two binary variables. See more here. | 
| Cost-Sensitive Learning | An learning approach where each classification error is assigned its own weight to fine-tune the model’s predilection towards certain error type1 2. | 
| Equivocal Zone | An interval of continuous output in which no binary class label is assigned. | 
| Applicability Assessment | The determination of whether a new input is similar enough to a model’s training data for a reliable prediction. | 
| Demographic Parity | A fairness criterion used to assess whether the outputs of a predictive model is independent of demographic groups (e.g. race, gender, or age). | 
| Predictive Parity | A fairness criterion used to assess whether the outputs of a model have equal positive and negative predictive values across demographic groups (e.g. race, gender, or age). | 
| Equalized Odds | A fairness criterion used to assess whether the outputs of a model have equal true positive rates (sensitivity) and false positive rates (specificity) across demographic groups (e.g. race, gender, or age). | 
| Global Explainability | The ability to estimate each feature’s impact on model outputs aggregated across an entire data set or feature space. | 
| Local Explainability | The ability to estimate each feature’s impact on model outputs for any given individual prediction. | 
| Governance | The processes by which organizational responsibilities and decisions are divided, evaluated, and executed. | 
| Deployment | Making a model or application accessible to other computers within a network. | 
| Production Environment | The software systems and infrastructure in which live applications are hosted and run for day-to-day operations. | 
| Development Environment | An isolated copy of the production environment where software changes can be tested without risk of impacting live operations. | 
| Application-Programming Interface (API) | A protocol or framework by which various software applications can communicate with each other and exchange data or predictions. | 
| Human-in-the-Loop | An implementation paradigm where model outputs are directed towards an expert user for incorporating into their decision-making before an action is taken. | 
| Data Drift | Divergence in input data away from the initial model training data set. | 
| Concept Drift | Divergence away from the training data in the target labels or context in which predictions are to be made. | 
| Continuous Integration (CI) | A DevOps principle in which changes to software are incorporate in small, manageable chunks continuously rather than large overhauls. | 
| Continuous Deployment (CD) | A DevOps principle in which updates to software a pushed to live environments without large periods of maintenance or down-time. | 
References
1. 
Ling CX, Sheng VS. Cost-Sensitive Learning [Internet]. In: Sammut C, Webb GI, editors. Boston, MA: Springer US; 2011. p. 231–5.Available from: https://link.springer.com/10.1007/978-0-387-30164-8_181
2. 
Mienye ID, Sun Y. Performance analysis of cost-sensitive learning methods with application to imbalanced medical data. Informatics in Medicine Unlocked [Internet] 2021;25:100690. Available from: https://linkinghub.elsevier.com/retrieve/pii/S235291482100174X