The CoreNLP tool [1] is running on the Kubernetes webservice platform and provides a natural language processing service to be used by other tools like platypus-qa and may be used to power research projects.
Because the ML models it uses are large (they take more than 1 GB each), that does not allow to have models for more than one language to be loaded in memory at the same time and so leads to loading times of ~5s when a request in an other language come in.
Is it possible to increase the memory limit of the corenlp container to something like 6GB? It would allow models for English, German, Spanish and French to be hopefully all in memory at the same time.