We currently have no easy way to run large scale analysis on the wikidata graph. WDQS and blazegraph are not suited for this scenario. Hadoop seems to be a better fit. Discussing with @JAllemandou we believe that a simple parquet file with quads might be sufficient for now.
Description
Description
Details
Details
Subject | Repo | Branch | Lines +/- | |
---|---|---|---|---|
Add WikidataTurtleDumpConverter to rdf-spark-tools | wikidata/query/rdf | master | +356 -10 |
Event Timeline
Comment Actions
Change 570324 had a related patch set uploaded (by DCausse; owner: Joal):
[wikidata/query/rdf@master] Add WikidataTurtleDumpConverter to rdf-spark-tools
Comment Actions
Change 570324 merged by jenkins-bot:
[wikidata/query/rdf@master] Add WikidataTurtleDumpConverter to rdf-spark-tools