As a machine learning engineer,
I want to deploy to production a model that detects peacock language by parsing the content of a paragraph and highlighting the problematic words.
I also want a reproducible way to generate and improve the model, making iteration easier and the process less error-prone by version-controlling the code and reviewing it through code reviews.
To achieve this, I’ll build an Airflow pipeline and deploy it on the ML airlfow instance. This pipeline will cover three basic steps:
• Dataset generation
• Model training
• Model evaluation
If not fully automated, the process of re-deploying a re-trained model from this pipeline should still be easy and well-documented.
Last but not least, an SLO dashboard using pyrra should be created for this service using the SLO template instructions.
- Model that detects peacock words for English
- Airflow pipeline for the model
- Dataset generation (fetching/preprocessing)
- notebook
- revert-based and recent newcomer' edits clean ver. notebook
- template-based dataset clean ver. notebook
- restructure the following notebooks
- move to Airflow
- try out the Airflow dev instance T391940#10792759
- create a config file with signals for top-20 wikis in T389445
- shift to Airflow ml instance
- notebook
- Model evaluation
- notebook
- move to Airflow
- figure out using WMFKubernetesPodOperator for model inference (and if it is production ready)
- Model training
- notebook
- restructure the training notebook, and consider retaining/fine-tuning workflow
- move to Airflow
- figure out using WMFKubernetesPodOperator for model training
- notebook
- Dataset generation (fetching/preprocessing)
- SLO dashboard T390706


