We're looking to host a big dataset on labs. We think it will be very useful to analysts who are working with PAWS and the tool developer community. See T145655
This dataset contains a row for every month that an article was visible in English Wikipedia. 5 million articles * ~100 months since 2002 = ~500 million rows. The output TSV file is 55GB uncompressed.
So now the question for discussion: Can we load this into user DB on labs without making @yuvipanda or @jcrespo sad?