Page MenuHomePhabricator

Come up with a NSFW filter for images in NewcomersTasks
Open, Needs TriagePublic

Description

It seems to me that newcomers will be very surprised, sitting at work, with such content on their page. I know about the principle of disclaimer, but in our case it can be frustrating.
F34451984 (NSWF)

Event Timeline

Restricted Application added a subscriber: Aklapper. · View Herald Transcript

Thanks, @Iniquity. We've talked about this a little bit, but haven't done anything to filter out articles that might not be best as the first articles for newcomers to see. If there were a template or category that was on such pages, it would be easy to filter them out, but I don't know that such a uniform template/category exists. I understand the idea of NSFW filters on images, but in this case, I think what we would be trying to filter out are articles that may contain such images (or even surprising content that is not images). Maybe we can think about this more in the future.

Thanks, @Iniquity. We've talked about this a little bit, but haven't done anything to filter out articles that might not be best as the first articles for newcomers to see. If there were a template or category that was on such pages, it would be easy to filter them out, but I don't know that such a uniform template/category exists. I understand the idea of NSFW filters on images, but in this case, I think what we would be trying to filter out are articles that may contain such images (or even surprising content that is not images). Maybe we can think about this more in the future.

Yes, I agree that there may be inappropriate content other than images. For example, we have this template (https://ru.wikipedia.org/wiki/Шаблон:Ненормативная_лексика), which is placed on articles containing obscenities and obscene vocabulary. I might start a discussion about the same template for articles with nude images.

Getting local templates will be too difficult to deal with. It is already very difficult to find which templates we have to use for newcomers tasks on most wikis! The example given for Russian Wikipedia only has 6 interwiki links total for Wikipedia. If we decide to remove what's considered as NSFW options (which can vary from one human to another and lead to potential censorship), we need to find a centralized option.

Filtering out these images should be done from Commons, maybe using https://commons.wikimedia.org/wiki/MediaWiki:Gadget-NSFW.js like already suggested. If any NSFW image is in the article, then we could exclude the article.

Related: T264045: Not Safe for Work (NSFW) media Classifier for Wikimedia Commons.

Ideally, NSFW articles would be flagged in the search index, but that seems way beyond what's plausible for us to do.

Another, less ideal but more realistic option would be to store the image's NSFW state somewhere (MCR? page_props? T282585: Provide for asynchronously-available MediaWiki parser content fragments / components might be relevant although that's about properties derived from page content, and files are handled very differently), expose it in imaginfo API requests, and then we could filter on the client side like we do for protected articles.

All this is blocked on T279416: Deploy Image content filtration model for Wikimedia Commons, though (the NSFW gadget uses WDQS, I doubt we want to do that in production frontend code).