User Details
- User Since
- Nov 9 2016, 7:25 PM (473 w, 1 d)
- Availability
- Available
- LDAP User
- Unknown
- MediaWiki User
- Pintoch [ Global Accounts ]
Sat, Nov 29
Thanks @Tarrow for working on this!
Sep 21 2025
I didn't know about the https://github.com/toolforge org, I wonder about its scope and governance. Would it make sense to add all standards committee members there, perhaps?
Aug 3 2025
The Tool sweep is an initiative initiated by @Legoktm which attempted to identify license issues in toolforge tools (among other issues). I wonder if there is an easy way to get an overview of how many reviewed tools were found not to comply?
May 21 2025
The page no longer exists, so it can't be imported anymore.
The property now has more than 3,000 uses so I think we can consider this dataset as imported.
As this is rather old, I am going to tentatively close this. If the problem persists, I would recommend opening a ticket on the bug tracker of the reconciliation service instead:
https://gitlab.com/nfdi4culture/openrefine-reconciliation-services/openrefine-wikibase/-/issues
This is already quite some years old, so I don't think it makes sense to keep this open for other tools. If more tools are affected, we can open fresh tickets for them.
I'm closing this as I don't think this coordination can or should happen. People make different formats for representing edits, which correspond to different use cases, and there's no problem with that.
May 12 2025
May 4 2025
May 3 2025
Fantastic, thank you so much!
May 1 2025
I have replicated this issue by creating a new Wikitech account "PintochTestDeleteMe", associated it with a fresh SSH key as instructed in the tutorial and attempted the login with that test account, unsuccessfully.
I am still able to log in via my normal account ("Pintoch").
Apr 4 2025
Ah sorry, I hadn't registered the difference between the two types of sessions. Friday 3pm looks ideal!
Thanks! Maybe some time on Saturday would be nice? But happy with anything, really.
Apr 3 2025
Apr 2 2025
I could hardly agree more, and it has indeed been suggested in T203557 shortly after the launch of this tool in 2018.
Mar 27 2025
Likely caused by https://github.com/OpenRefine/OpenRefine/issues/7231.
Hello @Tarrow, I was happy to hear you on the Between the Brackets podcast, where you mentioned EditGroups, which reminded me of this ticket :)
Mar 3 2025
Maybe that's a good occasion to say (in a personal capacity) that it was a pleasure working with WMSE on this. I'm also very grateful to Sunil Natraj who tackled a lot of the priority issues identified on the roadmap created by WMSE.
Feb 26 2025
This is fantastic, thank you so much for offering your help! Is it okay if I follow up by email to coordinate your onboarding? We could have a quick call to walk through the codebase together, work through the process of making a small change to the tool and make sure you're able to deploy it.
Nov 18 2024
Nov 3 2024
Oct 15 2024
I've signed the NDA a few days ago, let me know if more is needed
Oct 1 2024
That's a move I have already done, and it did help with making the service more stable indeed.
The remaining issues I am aware of:
- Celery workers getting killed, probably because of memory usage, meaning that revert tasks get interrupted. I have played a bit with the concurrency and memory allocation settings to make this happen less often, but solving the root cause of the problem would be better
- A need to migrate out of EventStreams to recent changes polling (I guess?) to make it possible to deploy the tool on other wikis (for instance Wikibase.Cloud)
- Need for ongoing care for the tool to make sure it stays in sync with the wikis (if it stays out of sync for longer than what the EventStream covers, we miss edits)
- Various other issues as recorded on the issue tracker: https://github.com/Wikidata/editgroups/issues
Sep 30 2024
Sep 26 2024
The latest version of Django (5.1.1) only supports MariaDB 10.5 or higher, and ToolsDB is currently running 10.4.29, so that's a hurdle I encountered while keeping EditGroup's dependencies up to date. If others encounter the same issue, the latest version of Django to work with MariaDB 10.4.29 seems to be 5.0.8.
It would be great to upgrade :)
Sep 25 2024
Sorry about that! I thought it could be interesting to get a better overview of the urgency of tasks in this board by trying to get a sense of how useful and relied on the mentioned tools are, but if that's not the right way of doing things, I'm happy to step back.
Sep 23 2024
Quoting the Q2-Q3 usability research :
Sep 19 2024
I'm tagging this as low priority given that there are alternatives such as https://tools.wmflabs.org/event-streams as mentioned above.
@waldyrious hello! I'm looking at the backlog of tools needing maintainers in https://phabricator.wikimedia.org/project/view/2952/ and found this.
I don't know what the original tool did, but given that there is since a well-established query builder at https://query.wikidata.org/, I am not sure it is still worth trying to resurrect the old tool just for the sake of providing input for UX research (but maybe it had great design ideas…).
Given that you removed your assignment, I would be inclined to close this task, no?
Sep 17 2024
Sep 8 2024
The OpenRefine developer in question was me ^^ (surprise surprise!)
Aug 8 2024
It would be interesting to have such an API, for instance for mass uploading tools like OpenRefine. This could potentially enable us to check for close duplicates before upload.
It would be ideal if the hashing algorithm used could be sort of standardized, meaning that it could also be implemented client side. But I understand if this is too much of a stability guarantee to offer.
Aug 2 2024
There will also be a meetup in the Belgrade room on Friday, 18:00-19:00:
https://wikimania.wikimedia.org/wiki/2024:Meetups/OpenRefine
Jul 15 2024
Jul 8 2024
Jun 14 2024
Ok, I wonder if it could be linked to the fact that I tried creating the redis job earlier, back when the quota didn't allow for it? I'm not sure.
Jun 13 2024
I have tried the approach described at https://wikitech.wikimedia.org/wiki/Tool:Containers#Redis_container and it works! The editgroups tool is functional again thanks to that. High five!
Jun 11 2024
@Jason.nlw I believe @Gnoeee is right: you should be able to create the tags yourself as administrator of the Wikibase.Cloud instance.
Jun 9 2024
May 23 2024
@fnegri I access the logs like this:
become editgroups kubectl get pods # find the pod id that starts with `editgroups.celery.sh` kubectl logs editgroups.celery.sh-58496b6654-gp5pj
In the current logs there seems to be a single occurrence of this error (the others such as "Worker exited prematurely" are likely unrelated, but not sure).
May 22 2024
Quick update on this: undoing edit batches on Wikidata via the current EditGroups has been broken for about two months now and the few hours I have spent trying to fix it have not been sufficient so far.
I can also observe this in the editgroups tool, which uses Celery:
[2024-05-22 13:39:42,965: WARNING/MainProcess] consumer: Connection to broker lost. Trying to re-establish the connection...
Traceback (most recent call last):
File "/data/project/editgroups/www/python/venv/lib/python3.11/site-packages/redis/connection.py", line 706, in send_packed_command
sendall(self._sock, item)
File "/data/project/editgroups/www/python/venv/lib/python3.11/site-packages/redis/_compat.py", line 9, in sendall
return sock.sendall(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
TimeoutError: [Errno 110] Connection timed outFeb 28 2024
Jan 28 2024
I added some specific instructions to setup Django to use a user database:
https://wikitech.wikimedia.org/wiki/Help:Toolforge/My_first_Django_OAuth_tool#Using_MariaDB_instead_of_SQLite
Those are also documented independently here (which I improved a little):
https://wikitech.wikimedia.org/wiki/Help:Toolforge/Database#Python:_Django
Oct 31 2023
Ok, I think I still don't understand it fully, but I trust you do and I won't stand in the way ^^
Oct 30 2023
For OpenRefine(and other tools) the benefit would be that RDF ontologies can be extended very easily so that tools can define their own (namespaced) properties.
Thanks for reviving this thread I had forgotten about ^^
Sep 28 2023
@Jason.nlw it is possible that the issue is on the OpenRefine side and not on the Wikibase.Cloud (not sure - I haven't checked!).
What is the preferred way to listen to the updates of a Wikibase.Cloud instance? (For instance, how is the Query Service updated?)
Currently EditGroups only supports listening to updates via the Wikimedia Event Service, which is obviously only available for Wikimedia projects and not for Wikibase.Cloud. Porting it to other Wikibase instances would likely involve adding support for updating via Recent Changes polling.
There is a ticket about this:
https://github.com/Wikidata/editgroups/issues/5
Sep 14 2023
Thank you @DannyS712! Let's say this patch solves the problem :)
Aug 2 2023
Yes, if it was unclear from my comments I can try to clarify again here. From an API user perspective my preferences are (from most preferred to least preferred):
- wikibase-entityid datavalue type, for consistency with the targeted user experience (EntitySchemas being entities themselves)
- string datavalue type, which is inconsistent with the UX but has the benefit of being an established datavalue type, which any existing API client library is bound to support already (Wikidata-Toolkit, Wikibase-SDK, pywikibot, …)
- A new datavalue type, which will likely require some light changes in most API client libraries, and perhaps lead to failures of various severities until those changes are made (for those which have not anticipated the introduction of new datavalue types)
Jul 19 2023
@Addshore wrote a blog post summarizing the options around this problem and I think it's a very worthy read:
https://addshore.com/2023/07/wikibase-and-reconciliation/
Jul 16 2023
May 11 2023
It's not just useful for testing purposes. For applications like OpenRefine, which normally run on the user's machine directly and are not meant to be hosted, it is important that the callback can be a localhost URL, therefore using HTTP. OpenRefine itself runs as a locally hosted web app (typically at http://localhost:3333/).
May 4 2023
If it's not too difficult, it would be helpful to have that behavior in list=recentchanges too, typically for tools which do recent changes polling.
(EditGroups sadly uses the EventStreams service and not RC polling directly - not sure if changing list=recentchanges would also impact the EventStreams as well?)
Thank you both, this should simplify the EditGroups tool quite a bit! Here is an example of page where I had to render entity links manually, with some Javascript post-processing code:
https://editgroups.toolforge.org/b/CB/3683873dde8d/. (I guess I will keep this logic for a while, for all the revisions fetched already, but it could be dropped in some years)
Mar 1 2023
Thanks for the heads up, I have opened an issue about it on OpenRefine's side: https://github.com/OpenRefine/OpenRefine/issues/5658
Jan 24 2023
Agreed, I would say PAWS replaces this task.
Jan 5 2023
Here is the current status of this issue:
Dec 10 2022
A feature request came in: handling changes of username on the wiki. I suspect this is a feature that would likely come "for free" in any reimplementation of the current tool as a MediaWiki extension, because it would rely on the existing SQL tables in MediaWiki to represent users.
Aug 31 2022
This works fine when using OpenRefine locally, so it probably has something to do with the deployment indeed. I have not checked, but I suspect the frontend code redirects the user to the project page using some relative URL, which fails in the PAWS context.
I do not know if this can easily be done, but I would assume one radical way to ensure that this does not happen would be to host OpenRefine at the root URL of some automatically-generated domain (instead of using a subpath).
If there are sensible fixes that can be applied on OpenRefine's side to make this easier, we can totally consider them.
Aug 16 2022
hello all, it's just caused by the change of format for our download URL. I have added a suggested change to the pull request that should hopefully fix it:
https://github.com/toolforge/paws/pull/187#pullrequestreview-1074315167
Jul 19 2022
Jul 12 2022
Sounds good! I think @Spinster also intends to write a few follow-up tickets.
Jul 9 2022
Should we close this as done?
Jul 5 2022
This bug was mentioned today at LSWT'22 by @SebastianHellmann as a blocker for integrating Wikidata in linked data applications.
Jun 23 2022
Jun 19 2022
May 21 2022
Thanks to @matthiasmullie we now know that this API response cannot return the pageid, because it is not known when the API response is sent back. The pageid is allocated asynchronously afterwards.
May 8 2022
However, wouldn't it be even simpler for the user to not even need the page id, and be able to add claims with a Commons file name?
Apr 28 2022
One helpful step in that direction would be to return the page id in the JSON response of a successful file upload. That should be fairly straightforward and would bring the number of requests required from three to two.
Apr 2 2022
I have not checked but I trust you!
