Thu, Feb 16
Wed, Feb 15
Tue, Feb 14
Per T128806: Switch German Wikipedia to uca-de category collation, this would need some consensus. I'm going to merge this task into that one.
Mon, Feb 13
Thank you, @Stevietheman. Would it simplify things if we defaulted to 1000 for every project without there being a config to change it? It isn't computationally expensive as it was with the prior bot.
I have a question about the current config. What's the purpose of "Number of pages"?
Sat, Feb 11
Fri, Feb 10
I'm not handling the database at all, at the moment. It generates report on the fly and puts it on the wiki page. See https://en.wikipedia.org/wiki/User:NiharikaKohli/Test_popular_pages for example.
Sat, Feb 4
Thanks for reporting, @Shizhao. T156198: Special:PageAssessments fails to load on English Wikipedia seems related.
I think it's safe to close this task as done now. CC @DannyH
Fri, Feb 3
More problematically, there are various issues with how links with raw unicode fragments behave when used outside the browser. For example, you want to share the article and copy the text from the address bar to an email - will the recipient's email client recognize where that link ends? Say we want to add a feature that shares section links to Twitter - will Twitter correctly identify where the link ends or break it at the first comma or such? What happens when that link goes through some system that is not Unicode-aware? What happens if the fragment contains a <script> tag - are all the places where we put links ready for handling that securely?
Is this still happening?
The tool is stuck on the same problem that happened a while ago. It keeps reconnecting to IRC if it doesn't get any responses for 1000 seconds. See this commit.
I think it's time we handle this properly. According to @eranroz:
I think it is not good to always get it to higher number, but either to completely remove it/adjust it per wiki/get a reliable way to validate we dont miss updates (such as querying the SQL revision table as an alternative).
Adjusting it per-wiki (as a param in the cronjob) might be the easiest solution at the moment. This would still need a bit of refactoring in Eranbot.
That's perfect. :)
एक समस्या यह पेज लोड किया गया था: कैलिफ़ोर्निया doesn't make sense. :)
Thu, Feb 2
I did enable the cronjob in Eranbot but it's not collecting any data. Will require some deeper analysis.
The pageviews API does not yet have the data for Wikiproject pages. According to T141010: Adding top counts for wiki projects (ex: WikiProject:Medicine) to pageview API this is probably going to happen this quarter
Is this ticket talking about just writing the bot which writes given content to a given page or also compiling the report? How are we getting that data?
This ticket has taught me so much. :)
Wed, Feb 1
Pinging @Marostegui in case he has time to look into this.
From my testing Safari tries to find the percent-decoded anchor first, if not found, it looks for the encoded anchor.
i.e. in https://tgr.github.io/scratchpad/unicode-fragment.html if you remove the id for 2. id="%E0%A4%AC%E0%A4%BE%E0%A4%A6%E0%A4%B2", it gives expected results for the first two links: 1, 1.
@kaldari Removed WIP. It's waiting for a review before merge.
Mon, Jan 30
Hi @MarcoAurelio, could you please create a page similar to https://en.wikipedia.org/wiki/User:EranBot/Copyright/User_whitelist on eswiki? It lists trusted users whose edits don't appear on CopyPatrol. I tried creating the page (with only the first line of text "Add usernames alphabetically, one per line, with the format * [[User:Example]]") but it didn't let me create the page for some reason.
Fri, Jan 27
Could we try asking on the village pump for olowiki about what collation they prefer?
Thanks. I fixed a bug where it didn't actually wait for the user to confirm. Also sadly, it's restricted to watching/unwatching top 50 pages at the moment. See https://www.mediawiki.org/wiki/API:Watch
It's possible to make it 500 if the Ajax request happens by a bot, which should be possible to do with a simple tool labs service, if there's any interest. :)
Wed, Jan 25
Tue, Jan 24
@Legoktm, could you shed some light on shadow namespaces project status? :)
We'd like to help, if possible.
Bump. There's no media upload dialog as tested with Chrome and Safari.
Mon, Jan 23
I encountered this today while trying to clone a repo.
vagrant@mediawiki-vagrant:/vagrant/mediawiki/extensions$ git clone https://gerrit.wikimedia.org/r/operations/mediawiki-config Cloning into 'mediawiki-config'... remote: Counting objects: 3902, done remote: Finding sources: 100% (1676/1676) remote: Getting sizes: 100% (740/740) remote: Compressing objects: 99% (20024/20025) error: RPC failed; result=56, HTTP code = 200MiB | 74.00 KiB/s fatal: The remote end hung up unexpectedly fatal: early EOF fatal: index-pack failed
@MarcoAurelio The script just finished running. Let me know if something seems amiss.
The way CopyPatrol currently works on Wikipedia(s) is:
- It grabs the text added during an edit.
- Passes the text to Ithenticate to check for plagiarism.
- Makes it show up on the interface if it is indicated as plagiarized.
@MarcoAurelio Sure, I've scheduled it for deployment later today. Will run the script once it's merged. :)
Hi, I wrote a simple user script to do this: https://en.wikipedia.org/wiki/User:NKohli_(WMF)/megawatch.js
@MarcoAurelio Thanks. Looks good to me. I can take care of running the maintenance script after the patch is deployed.
Sat, Jan 21
Jan 19 2017
@Legoktm Do you have any thoughts about disabling login on loginwiki altogether?
Jan 18 2017
Moving back to sprint so I don't forget to do this.
Jan 17 2017
Jan 13 2017
Jan 6 2017
Hi @Shoichi is the translation work currently in progress?
Wikia appears to be using https://www.mediawiki.org/wiki/Extension:SyntaxHighlight
Jan 5 2017
- Would it better to work off of the existing code in GitHub or start from scratch?
- I spent a good while trying to understand the existing code but it's beyond comprehension, pretty much. Without any documentation, it's almost impossible to understand what's going on. As far as I know, the bot uses dumps to create and populate tables which are then queried for the reports.
- Does Mr.Z-man want to be involved in this project?
- Per Ryan's comment above, yes, to some extent.
- Should we use the new Pageviews API to get the pageview data?
- This seems to be the logical way to do it. The pageviews API does not yet have the data for Wikiproject pages. According to T141010: Adding top counts for wiki projects (ex: WikiProject:Medicine) to pageview API this is probably going to happen this quarter. A comment in the ticket says that this will probably give back total views for top 1000 pages or so, which might not be perfect for our needs.
- Using the API will be more accurate than the dumps, most likely.
- A comment also mentions that the way this will work is going by how enwiki names its Wikiprojects, for example, "Wikiproject:Medicine". This might cause an issue with how other projects title their wikiprojects. I am wondering if there's a way to standardize the term "Wikiproject" across projects much like how "User" or "Template" are. Maybe a way to do this is using wikidata (Wikimedia Portal)
- Using the API will also provide us with more granular data - desktop and mobile views.
- Should we use PageAssessments for the assessment data? If so, do any changes need to be made to PageAssessments to have feature parity with the existing reports?
- We probably should. The PageAssessments API and database could come in useful. Besides the point made in T153790#2897180 I couldn't find anything else that'd need to change.
- What improvements can be made to the tool (besides fixing it)?
- Availability for all projects (not just enwiki)
- Granular data (mobile view stats)
- Data over a longer period of time, a year perhaps depending on whether pageviews API can provide that data (Requested on the author's talk page)
Jan 3 2017
Dec 23 2016
- Primary ticket is T11838: Send notification to account owner on multiple unsuccessful login attempts
- Extension Bawolff has worked on: https://www.mediawiki.org/wiki/Extension:LoginNotify
- Integrates nicely with Echo and optionally with CheckUser (to determine if the login is from an IP address you don't normally use)
- Uses IP caching and/or CheckUser to detect if login is from a known IP.
- Things we can work on:
- The security review (T140167) created T151414: LoginNotify cleanup which lists a bunch of outstanding things that need to be done before the extension is deployable.
- T135270: Update LoginNotify for AuthManager outlines a couple of hooks that need to be updated.
- The code currently has 11 FIXME's and 7 TODO's which could definitely be investigated and worked on to make the code more reliant, long term.
- Things to think about:
- The extension currently works on per-wiki basis. It would be a good idea to think about making it work with CentralAuth to detect login attempts on any of the wikis the user is attached to. Though the wishlist proposal doesn't mention it, it seems like an implicit given.