Oh hi. Nice to see you here.
User Details
- User Since
- Dec 11 2018, 9:39 PM (358 w, 1 d)
- Availability
- Available
- IRC Nick
- sukhe
- LDAP User
- Unknown
- MediaWiki User
- SSingh (WMF) [ Global Accounts ]
Yesterday
Sorry about this, this should now be fixed. And glad to see that Traffic was added automatically, thanks to @bd808 and @Ladsgroup for their work on this!
This is because of:
Tue, Oct 21
Hi @CRoslof: This is another ticket that we would like to take up and will need your help with so that we can reflect it in downstream services as well. Let me know if I should create a Zendesk thread for tracking by other Legal members? Thanks a lot of for bearing with us and helping us clean the ownership.
It looks like spicerack should check that alerts for the downtimed host have been resolved (not in firing state) before deleting the silence/downtime with ALERTS{alertstate="firing", instance=~"cp5018:.*"}
Mon, Oct 20
Thanks for filing this task! I think this is a good idea to reduce the manual updates to this list, and something we have failed to keep updated. We will triage this after discussion in Traffic.
Fri, Oct 17
Nice job indeed in pursuing this over the years, Brett!
[Adding Raine @kamila as well.]
Thanks for filling the task, @JVanderhoop-WMF. As per the discussion on Slack, the above sounds good.
Thu, Oct 16
Thanks to @Jhancock.wm for the help with this!
FWIW doing one or two hosts is more than enough. We will reimage them again anyway so it doesn't make sense IMO for you both to spend time upgrading all of them to trixie. If one or two reimage fine, please leave the rest to us.
Wed, Oct 15
I was also looped into a new request today. As part of the birthday initiative, the Fundraising team is developing a customized donation portal under the donate.wiki domain. Would it be possible to set up a redirect for this new portal as well? I don’t have the final destination URL yet, but we’d like to create the domain donate.wikipedia25.org to redirect to the donation portal once it’s ready. Is this something you could help with too?
Tue, Oct 14
Thanks for working on this! We will try our best to follow up on our end in making sure that Puppet is not broken on the cache hosts in Beta.
FWIW we have typically reimaged for this in the past. I am not suggesting, just sharing! And given that this is lvs1020, that might be OK? (Leaving to you both for the final decision.)
Thanks @MoritzMuehlenhoff for working on this and researching it. I am closing this for the reason mentioned above.
Cross-posting the comment from T405102#11273708,
Traffic discussed this in the team meeting today. We decided that given the above blocker, we should simply move to trixie and use OpenSSL (3.5.0) as shipped by trixie. The reasons for doing so are this: it doesn't make sense to upgrade to bookworm as that ships OpenSSL 3.0 and we have concerns around the performance (we haven't evaluated that as we have with 3.5 but we believe that to be the case from what we have seen).
Traffic has a dependency on Data Engineering for this to be executed, as we will need their help to run the query and generate the data set. That being said, we will make sure that we take this on when Data Engineering is ready (including doing the leg work).
(Is there anything -- including input -- required from Traffic on this? I am asking since we were added and can triage the task accordingly.)
Thu, Oct 9
Rolled out.
Wed, Oct 8
Thanks to a review by @bking, we have merged this. I think we can consider this as resolved. Thanks @LSobanski and @Gehel.
Thanks @MoritzMuehlenhoff, that sounds like a good plan to me but leaving to @CDobbins for the final word.
Tue, Oct 7
@bd808 and I discussed this today and decided to split this up in two parts:
hcaptcha200[1-2].wikimedia.org are ready.
Mon, Oct 6
Hi @akosiaris: Following up on this after a discussion during Traffic's planning with @Vgutierrez, and on behalf of the team.
Thanks for following up @cmooney. Other than the fact that we owed you a response on your last comment and never did -- that's on me of course -- yes, I think we can close it.
Fri, Oct 3
We will be resuming work on this in Q3 2025-2026.
@Shisma: Hi, this still needs a Wikimedia affiliate approval as per https://lists.wikimedia.org/pipermail/maps-l/2020-August/001729.html. Can you provide one, and after which we will need to check the technical feasibility?
Can someone help me double-check if this is still a problem? I don't see it in the dashboards above, selecting a more recent time interval.
Thu, Oct 2
@BCornwall from Traffic will be working on this, thanks!
Wed, Oct 1
@BCornwall from Traffic is handling this and will sync with @JAbrams and @jrbs for when the change is made live, to ensure incoming and outgoing email works as expected.
Tue, Sep 30
Mon, Sep 29
Note that @KOfori is out, this should be directed to @Kappakayala in the meantime.
I don't think this has anything to do with the UA. Is the command correct? Maybe try downloading via curl/wget and then piping to gpg --import '. curl https://www.mediawiki.org/keys/keys.txt | gpg --import - works for me for example.
[Removing Traffic since this is a MW issue. Please re-add if you disagree.]
I am curious: should we keep this open or should this be resolved now given that we have requestctl.wikimedia.org and that takes care of most of the pain points?
This was merged upstream in 9.2.x so we have inherited this change. Since we have not revisited this since 2021 but can pursue if required, I am taking the liberty to mark this as resolved.
We have had this for a while and the responses are padded. Marking as resolved.
What is the update on this, given that it has been a while and I am a bit confused reading the text and trying to follow the CRs.
And to be clear, by that I mean that this change is better suited for MW and not the CDN.
This is being moved on the Traffic workboard to "Radar/Not for service" as I don't think there is anything on our end to do here. Please let me know if you disagree and adjust accordingly.
My two cents: I think like you mention above, is this really required? I think this may be more work and doesn't give us a tangible benefit? Put differently I guess, what is the concern with multiple CRs?
Awaiting a response from @Lupascriptix before we can triage it again.
Paging has been disabled on this alert since July 2025. But even other than that, looking at the 90-day view, this has not happened in a while.
OK thank you. I am marking this as resolved for now. We can re-open as required.
Commenting from Traffic's side: this is in some ways, a trivial patch for us because we are simply setting an additional header. The challenge here, though, is understanding the header itself and the associated ramifications of setting it and also keeping it updated. For that, the Security should be/needs to be consulted, so this patch currently blocks on that happening.
The new cp hosts in codfw, Intel(R) Xeon(R) 6730P, support QAT, so we can experiment there and see if that helps.
This is worth investigating/researching but I am marking this as "Low" and we can revisit it for Q3 2025-2026.
We have made progress in T301605, and specific to this task, we ramped up traffic to magru already.
There has been no follow-up on this after Jun 2024. @MatthewVernon: should we keep this open?
Because of the regressions observed in OpenSSL 3.x, we never finished the upgrade for the cp hosts to bookworm. This task is stalled while we go through the possible upgrade paths for either OpenSSL, or an upgrade to trixie itself, depending on the former.
This has been open for a while and there hasn't been any follow up from either side. @MGChecker: Please re-open if this issue still persists for you. Thanks!
@Fabfur: Is there any follow-up to do here or can we close this given the issue at hand was resolved? Note that in the recent switchover, we didn't observe this issue.
Thu, Sep 25
Wed, Sep 24
OK great, resolving this for now but yeah, let's add this to the docs and see how it serves us next time. Thanks and sorry for the delay. Adding Traffic to the tags as Nat has done now should prevent this from taking a week next time! :)