Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 35 Next »

Meeting Minutes and Project Files

Our Publication References

Vladimir A. Makarov, Terry Stouch, Brandon Allgood, Chris D. Willis, Nick Lynch, "Best practices for artificial intelligence in life sciences research", Drug Discovery Today, Volume 26, Issue 5, 2021, Pages 1107-1110, ISSN 1359-6446, https://doi.org/10.1016/j.drudis.2021.01.017.  Free pre-print available at https://osf.io/eqm9j/
Abstract: We describe 11 best practices for the successful use of artificial intelligence and machine learning in pharmaceutical and biotechnology research at the data, technology and organizational management levels. 

Summary of Our Paper in Drug Discovery Today ("The 11 Commandments")

  1. DS is not enough: domain knowledge is a must
  2. Quality data (…and metadata) à quality models
  3. Long-term data management methodology, e.g. FAIR for life cycle planning for scientific data
  4. Publish model code, and testing and training data, sufficient for reproduction of research work, along with model results
  5. Use model management system
  6. Use ML methods fit for problem class
  7. Manage executive expectations
  8. Educate your colleagues – leaders in particular
  9. AI models + humans-in-the-loop = “AI-in-the-loop” (Chas Nelson invented the term)
  10. Experiment and fail fast if needed. A bad ML model that is quickly deemed worthless is better than a deceptive model
  11. Maintain an Innovation Center for moonshot-type technology programs (this COE is an example of one)


Walsh, I., Fishman, D., Garcia-Gasulla, D. et al. "DOME: recommendations for supervised machine learning validation in biology", Nature Methods 18, 1122–1127 (2021). https://doi.org/10.1038/s41592-021-01205-4

Abstract: DOME is a set of community-wide recommendations for reporting supervised machine learning–based analyses applied to biological studies. Broad adoption of these recommendations will help improve machine learning assessment and reproducibility.


Business Analysis

Our business analysis is based on enumeration of business personas and the challenges that these personas may face in their work with AI/ML technologies. The proposed Good Machine Learning Practices (GMLPs) answer these challenges.



Persona

Challenges

GMLPs that address these challenges

Author(s)

Reviewer(s)

Data Scientist

DS1: How do I pick a suitable data modeling technique?

1. Make recommendations for methods suitable for specific problem classes; or for auto-ML systems

2. DS should learn the application domain and/or work with domain experts



DS2: How do I store and version a model?

1. Recommend model versioning systems

2. Refer to best practices in ML Model Registration, Deployment, Monitoring

3. Best practices and regulatory standards in requirements management and model provenance

Christophe Chabbert


DS3: How do I tune the hyperparameter values?

  1. Hyperparameters in AI/ML models are used to drive the learning process, and as such, cannot be inferred from the training data. Examples include topology and size of the neural network, or the learning rate and batch size.

  2. The number of possible combinations of hyperparameter value choices can be large. Grid search may not work. Use Bayesian, evolutionary, or early stopping algorithms.

  3. Evaluate the model by using the n-fold cross-validation, aiming to preserve the statistical characteristics such as class distribution or value ranges in both training and test partitions.

  4. Evaluate the robustness of the model with the standard deviation or the confidence interval of the performance metric calculated for multiple values of the test data. For a quality model, the standard deviation of the performance metric should be low. Note the ranges of input values for which the prediction quality is high or low.

  5. It is highly recommended to assess the generalization performance of an AI/ML model on a data set completely independent from the data sets used in hyperparameter tuning.

Samuel Ramirez

DS4: How do I validate a model?

1. Refer to the DOME recommendations

2. Quickly fail models that are bad. A deceptive model is worse than a bad one

Fotis Psomopoulos


DS5: How to assure scalability of AI systems?


Elina Koletou


DS6: How should I acquire and manage data?

1. Recommend best practices from the FAIR Toolkit

2. Evaluate data for “fit for purpose”, in particular, for the metadata quality

3. Refer to best practices in Exploration, Cleaning, Normalizing, Feature Engineering, Scaling 

Natalja Kurbatova, Christophe Chabbert, Berenice Wulbrecht


DS7: Where do I store predictions?


Christophe Chabbert


DS8: How do I publish AI models?

Publish model code, and testing and training data, sufficient for reproduction of research work, along with model results. Make recommendations for an appropriate minimal information model

Fotis Psomopoulos


DS9: How to protect IP?




DS10: How should the applicability domain of the model be expressed?

Use Applicability Domain Analysis Samuel Ramirez

Model User

MS1: How should I manage the data? Includes data protection, versioning, labeling

1. Recommend best practices from the FAIR Toolkit

2. Evaluate data for “fit for purpose”, in particular, for the metadata quality

3. Refer to best practices in Exploration, Cleaning, Normalizing, Feature Engineering, Scaling

Natalja Kurbatova, Christophe Chabbert, Berenice Wulbrecht


MS2: How do I avoid bias?




MS3: How do I make sure the model produces sensible answers?

1. Set-up a “human-in-the-loop” system. Recommend tools for this, if they exist

2. Set-up business feedback mechanism for flagging model results that do not align with expectations

Chas Nelson


Architect

A1: How to assure scalability of AI systems? 


Christophe Chabbert


A2: How should I manage the data?

See above for Model User

Christophe Chabbert


A3: Continuous Integration/Delivery

1. Refer to best practices in DevOps

2. Automated model packaging for ease of production delivery

Elina Koletou


A4: How to assure performance (execution speed)?


Elina Koletou


Executive

E1: How do I learn about the costs and benefits of AI/ML technologies and the limits of possible?

Make recommendations for conferences, training materials, education products, review papers, and books. These must be updated on a frequent cadence



E2: How do I value and justify investments in AI?

1. Make recommendations for technology valuation methodologies

2. Provide business questions, goals, KPIs

3. Continuously review model performance against the business needs (goals, KPIs)

Prashant Natarajan

E3: How do I enable technology adoption?

1. Bring expertise together to assess compatibility of new technology with business architecture

2. Ensure sufficient knowledge transfer so users are comfortable with adoption and use the new approach

3. Have a sustainability plan so resources are available when needed

Prashant Natarajan


Common challenges across ALL personas:

  1. How do I develop and use AI/ML models in an ethical manner (privacy protection, ethical use, full disclosure, etc)?
  2. How do I develop and use AI/ML models in compliance with the government regulations?     

Common responsibility across multiple personas:

  1. Model improvement is a shared responsibility between the Model User, Architect, and Data Scientist personas.

Glossary notes: In the context of this document these terms have these meanings:

  1. Performance - metrics used to evaluate model quality in terms of sensitivity and specificity, AOC, or similar metrics, or execution speed.
  2. Validation - test to confirm performance of the model (as defined above); but should not be confused with “validation” in regulated systems.

Project members:


Fotis

Psomopoulos

CERTH

Brandon

Allgood

Valo Health

Christophe

Chabbert

Roche

Adrian

Schreyer

Exscientia

Elina

Koletou

Roche

Frederik

van der Broek

Elsevier

David

Wöhlert 

Elsevier

John

Overington

Exscientia

Loganathan

Kumarasamy

Zifo R&D

Neal

Dunkinson

Scibite

Simon

Thornber

GSK

Irene

Pak

BMS

Berenice

Wullbrecht

Ontoforce

Prashant

Natarajan

H2O.ai

Valerie

Morel

Ontoforce

Yvonna

Li

Roche

Silvio

Tosatto

Unipd.it

Natalja

Kurbatova

Zifo R&D

Mufis

Thalath

Zifo R&D

Niels

Van Beuningen

Vivenics

Ton

Van Daelen

BIOVIA (3ds)

Lucille

Valentine

Gliff.ai

Adrian

Fowkes

Lhasa

Chas

Nelson

Gliff.ai

Paolo

Simeone

Ontoforce

ChristophBernsBayer
MarkEarllSyngenta

Vladimir

Makarov

Pistoia Alliance

  • No labels