Some ideas for how to incorporate word2vec in our models, now that it's available in production:
- Delta between mean embedding of previous revision vs current for editquality
- Maybe the mean of words added
- Mean embedding for draftquality
- Mean for articlequality
- Could look at statement labels included in the edit comment for itemquality
Make sure that the embedding data is only loaded into memory once.