[This seems like an odd way to request information, but it appears to be what is desired.]
I would like to see the raw results of the benchmarks as they do not align with my benchmarking. I would also like to see how the benchmarking was set up and run.
Some of the conclusions in the benchmarking document do not appear to be correct. In particular, there is no reason to run the load process on the same machines that are used to run queries. This means that a query machine could have less memory. I wouldn't recommend less memory because a large-memory machine should be able to handle multiple query streams in addition to an update stream, but that doesn't mean that a machine with as little as 64GB of memory could not be an effective WDQS server.
None of the SPARQL engines are completely compliant with the SPARQL 1.1 specification. Quantifying non-compliance is thus a useful thing to do.
The query sets evaluated from WDbench are probably the least interesting. The other query sets pose much more of a test for SPARQL engines.
I believe that it is possible to load Wikidata into Blazegraph, provided that one has sufficient patience. The problem is that a complete load takes a long time and that there is a bug in Blazegraph that appears to be related to parallel access to the data structures during load that causes loading to fail sometimes.
The work done to convert SCHOLIA queries to standard SPARQL 1.1 should be useful to convert WDQS to standard SPARQL 1.1 in general.