Movatterモバイル変換


[0]ホーム

URL:


Keyboard Shortcuts

Thread View

  • j: Next unread message
  • k: Previous unread message
  • j a: Jump to all threads
  • j l: Jump to MailingList overview
List overview
Download

Wikitech-lOctober 2024

wikitech-l@lists.wikimedia.org
  • 43 participants
  • 43 discussions
Start a nNew thread
Hi all,we're movingirc.wikimedia.org to a new infrastructure [1]. Thisservice is unrelated to the IRC network (Libera) we use for real timediscussions. Insteadirc.wikimedia.org is an IRC service forbroadcasting recent changes events from public Wikimedia wikis to beused by various bots connected to per-wiki IRC channels.irc.wikimedia.org should not be used for any new bots (which shouldrather use Eventstreams[2]), but we still have various important botsrelying on the legacy IRC-based infrastructure.The current setup is full of technical debt and ultimately based on apatched version of a very old release of ircd-ratbox with a relayservice written in Python 2.We are replacing it with a modern standalone implementation in Python,which broadcasts IRC notifications in a format compatible to what iscurrently in use by the legacy setup:https://github.com/paravoid/ircstreamLast week during the SRE Infrastructure Foundations hackathon theircstream production setup was created. All our tests have beensuccessful, so on Thursday October 10 at 08:00 UTC we'll switch theirc.wikimedia.org DNS name to the new setup.No changes are needed to any bots, but if you run into any issuesafter the switch, please notify us in the #wikimedia-sre-foundationsIRC channel or leave a note athttps://phabricator.wikimedia.org/T376014.Cheers,Luca, Simon and MoritzFootnotes:[1]https://wikitech.wikimedia.org/wiki/Irc.wikimedia.org[2]https://wikitech.wikimedia.org/wiki/Event_Platform/EventStreams_HTTP_Service
4 3
0 0

23 Jul '25
Hi all,With MediaWiki at the WMF moving to Kubernetes, it's now time to startrunning manual maintenance scripts there. Any time you would previously SSHto a mwmaint host and run mwscript, follow these steps instead. The old waywill continue working for a little while, but it will be going away.What's familiar:Starting a maintenance script looks like this: rzl@deploy2002:~$ mwscript-k8s --comment="T341553" -- Version.php--wiki=enwikiAny options for the mwscript-k8s tool, as described below, go before the --.After the --, the first argument is the script name; everything else ispassed to the script. This is the same as you're used to passing tomwscript.What's different:- Run mwscript-k8s on a deployment host, not the maintenance host. Eitherdeployment host will work; your job will automatically run in whicheverdata center is active, so you no longer need to change hosts when there’s aswitchover.- You don't need a tmux. By default the tool launches your maintenancescript and exits immediately, without waiting for your job to finish. Ifyou log out of the deployment host, your job keeps running on theKubernetes cluster.- Kubernetes saves the maintenance script's output for seven days aftercompletion. By default, mwscript-k8s prints a kubectl command that you (oranyone else) can paste and run to monitor the output or save it to a file.- As a convenience, you can pass -f (--follow) to mwscript-k8s to immediatelybegin tailing the script output. If you like, you can do this inside a tmuxand keep the same workflow as before. Either way, you can safely disconnectand your script will continue running on Kubernetes. rzl@deploy2002:~$ mwscript-k8s -f -- Version.php --wiki=testwiki [...] MediaWiki version: 1.43.0-wmf.24 LTS (built: 22:35, 23 September 2024)- For scripts that take input on stdin, you can pass --attach tomwscript-k8s, either interactively or in a pipeline. rzl@deploy2002:~$ mwscript-k8s --attach -- shell.php --wiki=testwiki [...] Psy Shell v0.12.3 (PHP 7.4.33 — cli) by Justin Hileman > $wmgRealm = "production" > rzl@deploy2002:~$ cat example_url.txt | mwscript-k8s --attach --purgeList.php [...] Purging 1 urls Done!- Your maintenance script runs in a Docker container which will not outliveit, so it can't save persistent files to disk. Ensure your script logs itsimportant output to stdout, or persists it in a database or other remotestorage.- The --comment flag sets an optional (but encouraged) descriptive label,such as a task number.- Using standard kubectl commands[1][2], you can check the status, and viewthe output, of your running jobs or anyone else's. (Example: `kube_envmw-script codfw; kubectl get pod -l username=rzl`)[1]:https://wikitech.wikimedia.org/wiki/Kubernetes/Kubectl[2]:https://kubernetes.io/docs/reference/kubectl/quick-reference/What's not supported yet:- Maintenance scripts launched automatically on a timer. We're working onmigrating them -- for now, this is for one-off scripts launched by hand.- If your job is interrupted (e.g. by hardware problems), Kubernetes canautomatically move it to another machine and restart it, babysitting ituntil it completes. But we only want to do that if your job is safe torestart. So by default, if your job is interrupted, it will stay stoppeduntil you restart it yourself. Soon, we'll add an option to declare "thisis idempotent, please restart it as needed" and that design is recommendedfor new scripts.- No support yet for mwscriptwikiset, foreachwiki, foreachwikiindblist,etc, but we'll add similar functionality as flags to mwscript_k8s.Your feedback:Let me know by email or IRC, or on Phab (T341553<https://phabricator.wikimedia.org/T341553>). If mwscript-k8s doesn't workfor you, for now you can fall back to using the mwmaint hosts as before --but they will be going away. Please report any problems sooner rather thanlater, so that we can ensure the new system meets your needs before thathappens.Thanks,Reuven, for Service Ops SRE
9 23
0 0
Database dumps
by Byrial Jensen 17 Apr '25

17 Apr '25
Until some weeks agohttp://dumps.wikimedia.org/backup-index.html usedto show 4 dumps in progress at the same time. That meant that newdatabase dumps normally was available within about 3 weeks for alldatabases except for enwiki and maybe dewiki where the dump process dueto size took longer time.However the 4 dumps processes at one time become 3 some weeks ago. Andafter massive failures at June 4, only one dump has been in progress atthe same time. So at the current speed it will take several months tocome thru all dumps.Is it possible to speed up the process again using several dumpprocesses at the same time?Thank you,Byrial
3 2
0 0
User-Agent:
by Domas Mituzas 17 Apr '25

17 Apr '25
Hi!from now on specific per-bot/per-software/per-client User-Agent header is mandatory for contacting Wikimedia sites.Domas
19 61
0 0
API issue
by ovskmendov@proton.me 17 Apr '25

17 Apr '25
Whenever I access the API, I get "Wikimedia Error: Our servers are currently under maintenance or experiencing a technical problem." Sometimes if I wait a bit, I'll get a few good requests, but it will shortly go back to the previous message. My code was working just fine a few days ago and I didn't change anything.Sent with [Proton Mail](https://proton.me/) secure email.
7 12
0 0

17 Apr '25
Hoi,This is an inquiry from my friend in academia, researching about Wikipedia.He would like to know whether there's a way to acquire a list of templatesincluding external links. Here are some examples including external links.https://ja.wikipedia.org/wiki/Template:JOI/dochttps://ja.wikipedia.org/wiki/Template:Twitter/docSuch links are stored in externallinks.sql.gz, in an expanded form.When you want to check increase/decrease of linked domains in chronologicalorder through edit history, you have to check pages-meta-history1.xml etc.In a such case, traditional links and links by templates are mixed,Therefore, the latter ones (links by templates) should be expanded totraditional link forms.Sorry if what I am saying does not make sense.Thanks in advance,--Takashi Ota [[U:Takot]]
13 24
0 0
EBNF grammar project status?
by Steve Bennett 01 Apr '25

01 Apr '25
What's the status of the project to create a grammar for Wikitext in EBNF?There are two pages:http://meta.wikimedia.org/wiki/Wikitext_Metasyntaxhttp://www.mediawiki.org/wiki/Markup_specNothing seems to have happened since January this year. Also the comments onthe latter page seem to indicate a lack of clear goal: is this just a funproject, is it to improve the existing parser, or is it to facilititate anew parser? It's obviously a lot of work, so it needs to be of clearbenefit.Brion requested the grammar IIRC (and there's a comment to that effect athttp://bugzilla.wikimedia.org/show_bug.cgi?id=7), so I'm wondering what became of it.Is there still a goal of replacing the parser? Or is there some alternativeplan?Steve
26 217
0 0

25 Mar '25
Do your eyes glaze over when trying to make sense of Phabricator tasks with a long and troubled history of patches? Do you struggle to find the real people in the sea of automated bot comments?I've made myself a user style to compact the Gerritbot comments on Phabricator. You might like it too:https://userstyles.world/style/13140/wikimedia-phabricator-reduce-gerritbot…(See screenshots and installation instructions there.)-- Bartosz Dziewoński
8 12
0 0
Phabricator upstream shutdown
by Brian Wolff 21 Mar '25

21 Mar '25
It sounds like phabricator upstream is going away:https://admin.phacility.com/phame/post/view/11/phacility_is_winding_down_op…Just curious, are we planning to continue using it long term or move tosomething else?--Brian
10 12
0 0
Empowering Businesses with Innovative Digital Solutions
by consultancyservices393@gmail.com 15 Feb '25

15 Feb '25
MCS offers software consulting and development services, delivering digital transformation. Our services includes web application development, mobile application development, chatbot development, Data Science & AI Development Services, MVP Development, End-to-End IT Strategy & Consulting and implementation support, intuitive digital solutions, and leveraging the latest technologies.visit here:https://www.mayuraconsultancy.com
2 3
0 1
Results per page:

[8]ページ先頭

©2009-2025 Movatter.jp