Page MenuHomePhabricator

Move wiki-specific robots.txt out of the global file to Mediawiki:Robots.txt on specific wikis
Open, LowPublic

Description

https://en.wikipedia.org/robots.txt is kind of messy. There are some commented out lines. There are some lines that are clearly not applicable to en.wikipedia.org. We should clean it up at some point.

Event Timeline

MZMcBride raised the priority of this task from to Needs Triage.
MZMcBride updated the task description. (Show Details)
MZMcBride subscribed.
Glaisher subscribed.

Change 239403 had a related patch set uploaded (by Glaisher):
Remove redundant entries from robots.txt

https://gerrit.wikimedia.org/r/239403

Change 239403 merged by jenkins-bot:
Remove redundant entries from robots.txt

https://gerrit.wikimedia.org/r/239403

Change 240065 had a related patch set uploaded (by Dereckson):
Tidy robots.txt

https://gerrit.wikimedia.org/r/240065

A change has just been deployed. Not deeming this done, though.

A change has just been deployed. Not deeming this done, though.

Anybody knows what's left to do here?

A change has just been deployed. Not deeming this done, though.

@Jdforrester-WMF: Do you (or someone else? Who?) know what's left to do here?

A change has just been deployed. Not deeming this done, though.

@Jdforrester-WMF: Do you (or someone else? Who?) know what's left to do here?

I don't. Possibly the opener of the task may wish to do so?

https://en.wikipedia.org/robots.txt looks better, but we're still including lots of directives that aren't applicable to the English Wikipedia.

Now that we have https://en.wikipedia.org/wiki/MediaWiki:Robots.txt and friends, I think it would be nice to clean out the global robots.txt as much as possible. Whether we track this work using this task or a new task, I don't really care.

Izno renamed this task from Clean up robots.txt files to Move wiki-specific robots.txt out of the global file to Mediawiki:Robots.txt on specific wikis.Aug 23 2018, 8:13 PM

https://en.wikipedia.org/robots.txt looks better, but we're still including lots of directives that aren't applicable to the English Wikipedia.

Now that we have https://en.wikipedia.org/wiki/MediaWiki:Robots.txt and friends, I think it would be nice to clean out the global robots.txt as much as possible. Whether we track this work using this task or a new task, I don't really care.

I think that should be done as part of this task and I've updated the title accordingly. We need:

  1. A global interface admin to add the directives for the specific sites to those sites's robots.txt files;
  2. A developer to remove the directives for the specific sites from the global robots.txt.