The approved RFC for cleaner URLs of the form /wiki/PageName?action=history proposed a simple
Disallow: /wiki/*?
robots.txt rule. Nemo bis and Dantman point out inadequacies in this (see T14619#1192903, the RFC Talk page, and 2013 mailing list discussion), e.g.
Excluding default "dynamic" pages like the history from crawling makes sense, but reducing the availability of content more than what we do now is unwarranted and a huge cost ...
It would be helpful to identify here all the classes of URLs we do and do not want crawled.