In T354659, we did exploratory work showing that it is feasible to build a single model that works for many (>50) or even all languages.
However, the current architecture of the model is not suitable to easily maintain such a model; for example, the training pipeline using a bashscript is not suitable to automate the creation/updating of multilingual models.
In this task, we want to improve the pipelines for training and inference.
- developing an end-to-end training pipeline for the multilingual model; ideally with airflow
- using the pipeline to train one or few (>=5) models for all languages (i.e. figure out the best way to group languages)
revise the inference pipeline for the new model (keeping in mind the requirements to serve it via LiftWing later on)(this will be sketched out in a separate task)