It's still useful to have so we can parallelize reading massive dumps. Similar to XML dumps. (Each entity would be in one dump.)
Description
Description
| Status | Subtype | Assigned | Task | ||
|---|---|---|---|---|---|
| Open | None | T88728 Improve Wikimedia dumping infrastructure | |||
| Open | None | T88991 improve Wikidata dumps [tracking] | |||
| Open | None | T46581 Partial Wikidata dumps | |||
| Invalid | None | T285307 Create randomly split partial entity dumps |
Event Timeline
Comment Actions
We should think about splitting obvious classes (academic papers and astronomical objects) before going random with the rest.
Comment Actions
I'm closing this for now as we'll need to look at the topic more holistically and I believe a random split is probably last on the list of things we want to do before some other more meaningful splits.