Machine learning from development into production as a team

What about
• Dependencies?
• Reproducibility?
• Continuous integration?

Save the hustle with these simple practices

1/6
Usually you start with a Jupyter notebook to make them robust especially working as a team nbdev is a life saver

GitHub github.com/fastai/nbdevq

2/6
nbdev is a library that allows you to develop a python library in Jupyter Notebooks, putting all your code, tests and documentation in one place. That is: you now have a true literate programming environment, as envisioned by Donald Knuth back in 1983!

3/6
When multiple people work on the project dependencies and reproducibility can become a nightmare. Docker to the rescue!

Docker for Data Science — A Step by Step Guide
towardsdatascience.com/docker-for-dat…

With Docker, you have your ready-made environment at your fingertips.

4/6
Now you have multiple versions of your machine learning model? Also different Configs? How to track?

MLflow github.com/mlflow/mlflow an open source platform for the machine learning lifecycle

@MLflow

4/6
Use MLflow for tracking your various models, experiments, results and config
Btw. if it is pure ML pipeline approach its also possible to do everything directly in MLflow and skip the Docker step

5/6
So how to bring your ML model into production, why not as a micro service with FastAPI?

Step by step guide towardsdatascience.com/how-to-deploy-…

@tiangolo

6/6

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Philip Vollet

Philip Vollet Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @philipvollet

19 Apr
Does BERT Pretrained on Clinical Notes Reveal Sensitive Data? • Large Transformers pretrained over clinical notes from Electronic Health Records (EHR) have afforded substantial gains in performance on predictive clinical tasks.

Paper arxiv.org/abs/2104.07762
GitHub

↓ 1/4
github.com/elehman16/expo…

The cost of training such models and the necessity of data access to do so is coupled with their utility motivates parameter sharing, i.e., the release of pretrained models such as ClinicalBERT.

2/4
While most efforts have used deidentified EHR, many researchers have access to large sets of sensitive, non-deidentified EHR with which they might train a BERT model (or similar).

Would it be safe to release the weights of such a model if they did?

3/4
Read 4 tweets
22 Jan
Just a normal machine learning paper: Can a Fruit Fly Learn Word Embeddings? Apple cider vinegar and some drops of dish soap → Is All You Need

Paper arxiv.org/abs/2101.06887

The mushroom body of the fruit fly brain is one of the best studied systems in neuroscience.

#nlproc
At its core it consists of a population of Kenyon cells, which receive inputs from multiple sensory modalities.

These cells are inhibited by the anterior paired lateral neuron, thus creating a sparse high dimensional representation of the inputs.
In this work they study a mathematical formalization of this network motif and apply it to learning the correlational structure between words and their context in a corpus of unstructured text, a common natural language processing (NLP) task.
Read 4 tweets
22 Jan
Hey NLP is not only machine learning it also has to do with language.

But you are not alone!

Here are some of my classics that deal with linguistics

Linguistic Fundamentals for Natural Language Processing
by @emilymbender

May all-time favorite & silver bullet! A must read.
The beauty of this book is that it explains the most important linguistic concepts in short chunks and illustrates this with easy to understand examples

Book amazon.de/-/en/Emily-M-B…

eBook morganclaypoolpublishers.com/catalog_Orig/p…
Speech and Language Processing by Dan Jurafsky and James H. Martin web.stanford.edu/~jurafsky/slp3/

You are pulled into the topics very immersively and you can not stop reading! You literally soak up the knowledge
Read 4 tweets
6 Jan
When deep learning meets causal inference: a computational framework for drug repurposing from real-world data - Drug repurposing is an effective strat to iden. new uses for existing drugs

Paper bit.ly/38kndHM
Blog bit.ly/398Zj11

#ai #Bioinformatics
Existing methods for drug repurposing that mainly focus on pre-clinical information may exist translational issues when applied to human beings.
Real world data (RWD), such as electronic health records and insurance claims, provide information on large cohorts of users for many drugs.
Read 7 tweets
4 Jan
Narratives: fMRI data for evaluating models of naturalistic language comprehension - MRI datasets collected while human subjects listened to naturalistic spoken stories.

Paper lnkd.in/gdHUhHy
Dataset lnkd.in/gSQFBpF
Datalad lnkd.in/gZBTz4T Image
The current release includes 345 subjects, 891 functional scans, and 27 diverse stories of varying duration totaling ~4.6 hours of unique stimuli (~43,000 words).
This data collection is well-suited for naturalistic neuroimaging analysis, and is intended to serve as a benchmark for models of language and narrative comprehension.
Read 4 tweets
27 Nov 20
A new SOTA open source semantic annotator ‘bbw’ for tabular data with the Wikidata knowledge graph.

GitHub github.com/UB-Mannheim/bbw

PyPI pypi.org/project/bbw

Notebook github.com/UB-Mannheim/bb…

$𝚙𝚒𝚙 𝚒𝚗𝚜𝚝𝚊𝚕𝚕 𝚋𝚋𝚠
Semantic annotation or tagging is the process of attaching to a text document or other unstructured content, metadata about concepts (e.g., people, places, organizations, products or topics) relevant to it.
Read 4 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!