The purpose of the spam blacklist is to prevent spam, not to prevent specific links from being added to a page. As bots are already vetted to ensure that they are not going to spam the wiki, they should be exempt from the spam blacklist. Among other issues this would solve (see the discussion in T36928 for more), archive bots will can attempt to archive a section containing a blacklisted link. The bot is unable to add the discussion to the archive page. This results in thousands of discussion being lost, usually without anyone knowing it happened.
What's the proposed implementation?
Bot is a default user group in mediawiki. Should code be added to SpamBlacklist to automatically add sboverride to all bot groups by default? Or do we need to do this one wiki at a time via config files?
I think, this right can be added to standard pack of bot's rights and can be removed in config of certain wikis by request of their users. Please, complete this request, because every month without this feature is inconvenient for bots and botmasters.
This should not be hardcoded like this, it should be coded as to allow additions of matches according to user rights, some should simply never be added (not even by admins, as the current SBL works, others should be disallowed only for unregistered editors or new editors. In this way one could even allow usage of link shorteners to e.g. extended confirmed users or users with ‘given’ rights, but disallow them being used by spambots (and then bots can replace all of them).
Note that overriding the SBL like this will result in massive vandalism fighting problems - if a vandalism edit removes the blacklisted link, and then a good follow up edit makes it impossible to revert the vandalism edit you cannot edit the vandalised information back in because you are not a bot.
@Beetstra can you elaborate on the usecase you care mentioning - is it this?
- A page exists, and has a SBL link on it
- A disruptive edit is made to the page, where the edit also removes the blacklisted link
- A constructive edit is made to the page
Please identify what is different then the prior state as well.
It has always worked this way, the change discussed here did not change this mechanism, did not make it worse, so this is not the place to make such claims. And secondly, you can roll back such edit, if you are rollbacker.
- Now try to revert the edit that vandalized … you can if the diffs are
not conflicting (that overrides the SBL), but if it is conflicting you
cannot, you have to edit it back in, but you cannot because that would add
a blacklisted link back in.
What I mean is, the only way to get the blacklisted link back in is through
a revert, undo, not through a regular edit. And situations do happen where
revert is wrong (you would revert a good, or multiple good, edit(s) which
you then have to re-do), and where undo does not work.
In short, even if blacklisted links on pages are not an immediate problem,
they ALL need to be removed or whitelisted. Allowing them being added (by
bots of all) is just plain wrong. There is a reason why blacklisted links
are blacklisted, most of them are crappy material that we do not want to
link to, and quite a significant number of them are blacklisted not because
they were spammed, but because the community did not want them to be linked.
Bots are not really "adding" blacklisted links, the links were already there. We can continue to block bots from doing their work, but the consequences IMO are much more severe than the edge scenario outlined above.
@Beetstra The reasons why we requested this change is
- because archiving bots can't archive discussion, if it contains a link, blacklisted after discussion. Ruwiki lost thousands of discussions because of it, it is one of the most disruptive events in Wikipedia history.
- because bots like AWB repairs broken link syntax on pages when edit in automatic mode, and if bot repairs spam link, it can't save edit.
@MBH: on en.wiki, bots just don’t archive when they can’t save. Morever, the discussions are not lost, barring oversight and deletion, everything is still there. The better solution is to break the link so it is disabled. Well, actually the better solution is to have the SBL be able to distinguish namespaces
Bots should not repair spam links, those links should be removed. There is a reason why stuff is blacklisted, the community does not want it. You want spammers to intelligently maim redirect links so a bot repairs it for them and they have their link? Spammers spam because it makes them money. Any link they get gives them a chance it gets followed.
Fix the SBL, not circumvent the problems it gives.
Yes, this is always like this, and that is why blacklisted links need to be removed, and certanly not en masse added by bots
Rollback and undo work indeed, but not everyone has the first right, and if there are follow
up edits, rollback does not work, and sometimes undo neither.