User Details
- User Since
- Jun 27 2020, 12:14 AM (298 w, 3 d)
- Roles
- Administrator
- Availability
- Available
- IRC Nick
- dancy
- LDAP User
- Ahmon Dancy
- MediaWiki User
- ADancy (WMF) [ Global Accounts ]
Thu, Mar 12
I just noticed this while doing local catalyst-api development.
Tue, Mar 10
@brennen I heard you were going to be looking into this stuff.
Wed, Mar 4
Mon, Mar 2
Notice to train operator @jeena: I merged https://gitlab.wikimedia.org/repos/releng/release/-/merge_requests/236 today. This removes the 5 minute sleep that occurs after a full image build (T390251) now that T412951 has been implemented. If this causes a problem, merge a revert commit for https://gitlab.wikimedia.org/repos/releng/release/-/merge_requests/236 to recover.
Thu, Feb 26
Tue, Feb 24
bookworm-php-sury and its dependent images have been updated.
Mon, Feb 23
I'm running into this problem today so I'll take this ticket.
Fri, Feb 20
Thu, Feb 19
Thanks everyone!
A burst of these errors happened on Feb 16 between 13:35 and 13:40. Here's a fresh trace:
Wed, Feb 18
Thanks for the work on this. I'm holding the train until the fix is backported.
Feb 13 2026
Feb 11 2026
Thanks @jnuche !
@elukey Congrats on the experiment!
Feb 10 2026
Feb 5 2026
I believe that the work done in T408740 has resolved this issue.
Feb 2 2026
Jan 29 2026
@Reedy I deployed a change to the kubernetes deployment monitor code in scap. Now it will not give up monitoring when it gets an error from kubectl.
Jan 28 2026
Jan 27 2026
Jan 26 2026
Here's a working .gitlab-ci.yml config using wmcs runners: https://gitlab.wikimedia.org/repos/m3api/tmp-m3api-oauth2/-/merge_requests/1/diffs
@LucasWerkmeister Thanks for setting up https://gitlab.wikimedia.org/repos/m3api/tmp-m3api-oauth2 and adding me as a member. I tried to run the CI pipeline of that repo on the main branch (using the normal D.O. runners) but it fails: https://gitlab.wikimedia.org/repos/m3api/tmp-m3api-oauth2/-/jobs/727347. I need a working pipeline as a starting point for debugging.
Jan 23 2026
Deployed via scap 4.235.0.
If a merge request description body consists of a single Bug: TXXXXX line, the link to Phabricator does not get created. I'm wondering if this is a new problem since the recent Gitlab UI changes. See the job link referenced in the task description for an example (https://gitlab.wikimedia.org/repos/abstract-wiki/wikifunctions/function-orchestrator/-/merge_requests/18).
Jan 22 2026
Jan 21 2026
Change deployed via scap 4.234.0
Jan 20 2026
@Clement_Goubert Please use scap lock --unlock-all --yes <unlock reason> (note: --bg is not passed here).
Thanks for the report @elukey. This sounds very promising!
Jan 16 2026
I confirmed that I did receive an email when I created a beta wiki account.
xref T412975
I ran sudo /usr/local/sbin/clean-stale-puppet-certs --clean on deployment-puppetserver-1.deployment-prep to take care of this.
Thanks for the work on this @thcipriani !
@jeena @thcipriani What's the status of the mediawiki-dev deployment-chart? I found that it is referenced in https://wikitech.wikimedia.org/wiki/Deployment_pipeline/Migration/Tutorial
Jan 15 2026
@Blake I've installed a new release of scap on the deploy servers. You can now use scap lock --all --bg and scap lock --unlock-all to achieve your goal. I recommend testing things out on deployment.eqiad.wmnet well in advance of the datacenter switch. Let me know if you have any issues.
Jan 14 2026
@A_smart_kitten patches are welcome! Let me know if you need help with getting something together.
Jan 13 2026
Potential ideas:
- Drop a lock file on the deployment server that scap detects, remove it at a later step
- Add a switch to scap lock --all that puts the lock on and returns, add a scap unlock --all command
Added
profile::tlsproxy::envoy::upstream_sni: null profile::tlsproxy::envoy::upstream_tls: false
to devtools project puppet according to @bd808's recommendation.
Looks similar to T414304
Jan 12 2026
I turned off this instance as part of T412975
Jan 9 2026
After reading the history of this ticket, it's not clear to me if the memory-optimized runners have been tried yet. Can someone try and confirm?
Jan 8 2026
Jan 7 2026
Jan 6 2026
@Andrew has since changed where he performing the build of the wikitech-static container image, so the main cause of this ticket has been resolved.