User Details
- User Since
- Nov 9 2015, 9:18 PM (454 w, 4 d)
- Availability
- Available
- IRC Nick
- gehel
- LDAP User
- Gehel
- MediaWiki User
- GLederrey (WMF) [ Global Accounts ]
Yesterday
Thu, Jul 25
Tue, Jul 23
Mon, Jul 22
Increase in traffic has been identified as coming from a single bot, which has been rate limited. Longer term solution is to add a pool counter (see attached patch).
Those 2 projects have not been used since forever as far as I know. I've archived them.
@hashar : it seems there is an issue with the Gitlab host key: https://gitlab.wikimedia.org/repos/maven/maven-test-project/-/merge_requests/1#note_95865. Could you check and ping @amastilovic if you need more context?
Fri, Jul 19
Thu, Jul 18
Not used by me or my team's. Unless @mpopov has an objection this can be deleted
Additionally, it would be useful to have some idea on the kind of queries we expect. For example: do we expect MoreLike requests to be based just on the article being viewed? Or do we want to further filter based on some user preference (their preferred topics for example)? This would help get an idea of the fragmentation of caching we can expect.
@brennen : Do you have an opinion on where to host artifacts coming from projects hosted on Gerrit? Does it make sense to create a project on Gitlab that would be dedicated to hosting artifacts generated outside of Gitlab?
Currently, only the Java 8 docker image supports open-ssh client. We'll need to replicate that functionality for other versions of Java, probably by moving that to the Maven base image. @hashar might be able to help.
Wed, Jul 17
Tue, Jul 16
Removing DPE SRE, this should be driven by Data Products.
It seems that the Gitlab model is to publish artifacts per project. At the group level, a read only package registry is available, which aggregates all the projects under that group. This works well for projects hosted on Gitlab, but asks the question of what do we do for projects hosted on Gerrit. We still need to publish them to Gitlab. Note that I don't fully understand how Gitlab works, I might be missing something.
Mon, Jul 15
The Search Plaform team will spend some time to investigate. There have been some issues with Dumps lately, which might have an influence here (the category graph is loaded from dumps). Another potential issue is that the category sub graph is too large in this case and we bail early for performance reasons (Deep Cat Search is a best effort service, that might not do an exhaustive search of categories).
Next step: brainstorming with 1 engineer from the web team
Option choosen: remore the delay / deduplication just for wikidata
Fri, Jul 12
This task is mostly done, but let's keep it open until we've published at least 1 project to Gitlab, with a CI workflow, to validate that this works outside of a simple test project uploaded manually.
Relevant comment on T367322#9965007