Movatterモバイル変換


[0]ホーム

URL:


Jump to content
Wikimedia Meta-Wiki
Search

Steward requests/Bot status

From Meta, a Wikimedia project coordination wiki
<Steward requests
For bot status on Meta-Wiki, see thebot policy andrequest page.
Requests and proposals
Steward requests (Bot status)
latest archive
Shortcut:
SRB
This page hosts requests for aBot flag on a Wikimedia wiki with no bureaucrats. Please read theinstructions before posting a request. Request to remove own permission should be done inSRP.

Please only make request hereafter gaining the on-wiki approval of your local community (i.e., don't list your bot here if you haven't consulted the local wiki for at least a week, or whatever local policies dictate).Previous requests arearchived.


Cross-wiki requests
Meta-Wiki requests

Instructions

[edit]

Before posting here

[edit]

Before posting here, please check the following points:

  1. If there is a bureaucrat on the local wiki (visit Special:ListUsers/bureaucrat on the wiki to check): Please ask them to fulfill your request usingSpecial:UserRights on the relevant wiki.
  2. Before requesting access here, make sure to obtain consensus on the local wiki if there is a community (see theindex of request pages). If there is no local community, request on this page with a note to that effect. You can ask local bot flag directly here forsome wikis in some cases (seeautomatic approval).
  3. Remember to read the local project's bot policy. This should be linked from thelocal bot request page; a few are listed below.DO NOT post your request here without having fulfilled the local policies first, and having waited for at least a week to gauge community opinion. In the case that the project does not have a local bot policy, the globalbot policy applies.
ProjectLanguages
Wikipedia

afanarastazbarbat-smgbclbebe-x-oldbgbnbrbscacecocsdadedveneoeseufafifrfrpfrrgaglhakhehihifhthuhyiaidieigioisitiujakakabkmkokrckshkukwladlblmolnltlvmap-bmsmgmhrminmrmsmtmymyvnapndsnenlnonvocospdcplpsptrmyrorusiskslsosrsusqsvszlthtotrttukuzvecvivowarwowuuyiyozeazhzh-classicalzh-min-nanzh-yue

Wikisource

deenhyitnoplptruvizh

Wiktionary

afarbrdeenesfrfygagdglidisitkakoltmsnlnoptrosvthtrukvowazh

Wikibooks

csdeenesfrfykoptzh

Wikinews

arcadeenesfafrptrutr

Wikiquote

enfafrhyitkaruzh

Wikiversity

ardeelenfritptruzh

Wikivoyage

deeneoesfritnlptsvzh

Wikimedia

CommonsMetaWikidataWikispeciesMediaWiki.org

Making a request

[edit]
  1. Place the following code at the bottom of the "Bot status requests" section below:
    === Bot name@xxproject ==={{sr-request| status=<!--don't change this line-->| domain=| user name=| discussion=}}(your remarks)~~~~

    You can use following form when applying for global bot status:

    === Global bot status for <yourbotname> ==={{sr-request| status=<!--don't change this line-->| domain= global<!--don't change this line-->| user name=| discussion=}}{{indent|''Not ending before{{subst:#time:H:i, j F Y (T)|+2 week}}''}}(your remarks)~~~~
  2. Fill in the values:
    • 'domain': the wiki's URL domain (like "ex.wikipedia" or "meta.wikimedia").
    • 'user name': the bot's user name (like "ExampleBot"). Leave it blank if you're requesting access for multiple bots.
    • 'discussion': a link to the local vote or discussion about the bot (for example, "[[ex:Wikipedia:Bots#ExampleBot]]").Notes: If there is no local vote or discussion, please say so below the template and explain why not. If there's no community,implement the standard bot policy instead.
  3. If anything is missing from your request, a steward will request more information.


Global bot status requests

[edit]

Global bots are enabled on most wikis using thestandard bot policy (seelist of global bot wikis andglobal bot access requirements). Request will be left open for at least 2 weeks to allow comments, and will be advertised viaMassMessage.

Please add requests at thebottom of this section using the template above.

Global bot status forUser:SchlurcherBot

[edit]
Status:   In progress
Not ending before 20:37, 3 November 2025 (UTC)

Requested task: Convert links fromhttp:// tohttps:// on all projects that allow global bots.

Links to relevant discussions on the task: For discussions in English Wikipedia see:

Edit mode: Fully automatic

Approval status: The bot is approved for this task in Commons (with ~4 million 6 million[✍️ 1] edits), DE-Wiki (with ~ 95'000 edits), EN-Wiki (with ~ 90'000 edits), PL-Wiki (with ~ 24'000 edits), FR-Wiki (with ~ 20'000 edits) and IT-Wiki (with ~ 15'000 edits), seeCentralAuth for an overview of contributions.

  1. Dušan Kreheľ, 2025-10-22 12:06 UTC
  2. Further details: Please seeUser:SchlurcherBot for details regarding programming language, code, function details, sources for pages, further technical comments and approval status.

    Comments:

    @مواطن تونسي: Hi, the checking is performed as part of the standard HttpClient functionality in C#. I've tested this behavior using expired or self-signed certificates (frombadssl.com). In these cases and as also seenhere, the HttpClient correctly throws an exception, and the exception thrown prevents replacement of the link by the bot. --Schlurcher (talk)06:56, 22 October 2025 (UTC)[reply]
    @Schlurcher Thank you for your feedback. I took the liberty of adding two URLs to the page in question to test the case where the site does not respond on port 443. Would it be possible to run one last test with the bot? --مواطن تونسي (talk)09:12, 22 October 2025 (UTC)[reply]
    @مواطن تونسي: I've run the test, an no updates were performed. The link to iort.gov.tn is checked and disregarded. The Wikipedia link is not checked at all. Please note that even though you added a http:// link to en.wikipedia, the rendered website automatically included a https href element in this case. As the bot picks up the href elements and does not process the page (also to handle templates that modify links and all kind of other scenarios), this is not a http link. You can also see this as a user, as if you hover over the link, the browser will show the https site as destination, at least in Edge. This is also reflected in the source code:
    <arel="nofollow"class="external free"href="http://www.iort.gov.tn/">http://www.iort.gov.tn/</a><aclass="external free"href="https://en.wikipedia.org/">http://en.wikipedia.org/</a>
    Hope this helps. --Schlurcher (talk)11:27, 22 October 2025 (UTC)[reply]
    Support Support following your feedback and tests. --مواطن تونسي (talk)11:36, 22 October 2025 (UTC)[reply]
    • Support Support however beware that just because a wiki allows global bots doesn't mean that they will allow it for your purpose without explicit authorisation. Many wikis, including the English Wikpedia, use the outdated 2008 rules that only allow global bots for interwiki redirects.Leaderboard (talk)07:04, 22 October 2025 (UTC)[reply]
    • Support Support --Dušan Kreheľ (talk)12:06, 22 October 2025 (UTC)[reply]
    • QuestionQuestion: Where is the complete source code? The user page of the bot links toc:User:SchlurcherBot/LinkChecker, but that only contains the HTTPS link extraction. I’d be interested in the entire process, for example how it matches links back to the wiki code that needs to be edited in the end. On huwiki, there are many infoboxes that contain things like[http://{{{website}}} Website] – will the bot get confused if it sees|website=example.org/page.html, without the protocol? Addinghttps:// at the beginning would break the link. On the other hand, templates like {{URL}} handle both URLs with and without protocol, so you could addhttps:// to their parameters. Is the bot smart enough to differentiate between these cases? If not, does it err on the side of not touching links? (By the way, where does it run: on your own computer or on a shared hosting like Toolforge?) —Tacsipacsi (talk)08:22, 26 October 2025 (UTC)[reply]
      @Tacsipacsi: I am currently unable to release the full code because as a hobby developer I have not fully tracked all individual contributions. Proper attribution is a requirement of the code pieces I've used, and without a complete record of authorship, I cannot ensure compliance. I appreciate your understanding. The LinkChecker was written 100% by me and thus made available. The entire process is as follows:
      1. The bot generates a parse of the page to be analyses and uses a custom re-write API request fromWikiDataDotNet for this
      2. It searches the html parse for all href elements and extracts the links. This is done viaHtml Agility Pack
      3. Links that are subsets of other links are filtered out to minimize search and replace errors via simple C# functions
      4. The bot checks if the identified http-links also occur in the wikitext, otherwise they are skipped. This is done via a simple string search
      5. The remaining links are handed over to the LinkChecker
      6. The final search-and-replace is done via a custom re-write and debugging ofDotNetWikiBot
      To answer your questions:|website=example.org/page.html will not confuse the bot, as the corresponding parsed link will be disregarded in step #4. I'm parsing the href elements and not the wikitext in #1 and #2 to avoid any such problems with templates that modify links (like archiving templates).
      All runs on the machine depicted oncommons:User:SchlurcherBot. This paticular task runs on the NVIDIA Jetson Nano 4GB, as the C# library to calculate the Levenshtein algorithm can be speed up significantly with the GPU power of the Jetson Nano. Are there any particular pages on huwiki that you want me to try? Perhu:Wikipédia:Bürokraták üzenőfala/Botjelentkezés limited tests are allowed prior to approval. --Schlurcher (talk)13:53, 26 October 2025 (UTC)[reply]
      Thanks for the response! Are you sure absolutely no other parts that could be published? WikiDataDotNet, Html Agility Pack and DotNetWikiBot are pieces of free software (the former two are MIT-licensed, the latter is GPLv2), so if the only copyrightable changes to them were written by you, you’re allowed (and welcome) to publish the modified source codes. I’d be especially interested in step 6, as that’s where an eventual bug can cause the most problems.

      Links that are subsets

      What does this mean? Could you share some examples?

      To answer your questions:|website=example.org/page.html will not confuse the bot, as the corresponding parsed link will be disregarded in step #4.

      Thanks, that’s good to know. It causes a few false negatives like the {{URL}} template, but false negatives are far less bad than false positives (in fact, false negatives are no worse than your bot not running at all).

      I'm parsing the href elements and not the wikitext in #1 and #2 to avoid any such problems with templates that modify links (like archiving templates).

      I realized there is another problem as well, which is hopefully also avoided by this: URL-like things that aren’t actually URLs, like XML namespaces. At least I hope that they aren’t linked.

      All runs on the machine depicted oncommons:User:SchlurcherBot. This paticular task runs on the NVIDIA Jetson Nano 4GB, as the C# library to calculate the Levenshtein algorithm can be speed up significantly with the GPU power of the Jetson Nano.

      I see. By the way, why is the text similarity check needed? That is, why do you consider pages that are not character-for-character equal (but very similar) equal?

      Are there any particular pages on huwiki that you want me to try? Perhu:Wikipédia:Bürokraták üzenőfala/Botjelentkezés limited tests are allowed prior to approval.

      No particular pages come to my mind at the moment. However, there is aproblem area I can think of: encoding. Would your bot findhttp://hu.wikipedia.org/wiki/Kezdőlap (if it wasn’t auto-fixed to HTTPS by MediaWiki)? Andhttp://hu.wikipedia.org/wiki/Kezd%C5%91lap? They are conceptually the same (and maybe the parsing you do actually makes them the same – I can’t tell this for sure without having seen the source code of steps 1 and 2), but they are written differently in wikitext. And what about the likewise equivalenthttp://példa.hu/ andhttp://xn--plda-bpa.hu/ (IDN; “példa” means “example” in Hungarian)? IDNs are not very common in Hungarian, but you might find some; and it’s probably much more common in non-Latin-script languages.
      By the way, why do you use different frameworks for getting the HTML and for editing (and even different languages – F# and C#)? —Tacsipacsi (talk)17:50, 26 October 2025 (UTC)[reply]
      @Tacsipacsi:Links that are subsets: This came from the enwiki bot request. Specially this edit:[1], where both the https archived and http non-archived link are given on the same page. I've now updated the initial filtering logic to correct for this: I'll now remove all http links that are contained in other links on the same page (checked against all http and https links). The revised edit with the updated logic is here:[2]. Generally the bot tries to be conservative. So if in doubt, it skips it.XML namespaces came up in the dewiki bot request and are not automatically linked by the Wikimedia software and thus not picked up by my bot.Text similarity check are needed as most commerical homepages use dynamic IDs for certain elements, like for ad containers to circumvent Ad Blockers. Also other pages, like archive.org for example, add rendering dates and last access dates into the source code dynamically. All these dynamic changes would invalidate a direct character match.Regarding IDNs: I've tested this a bit with non-latin-script languages and will continue to monitor this, but generall the Wikipedia software renders the href element ofhttp://xn--plda-bpa.hu/ tohttp://példa.hu/ and thus that link will be disregarded by check #4, which is an exact character string search. --Schlurcher (talk)19:12, 26 October 2025 (UTC)[reply]
    • Support Support but isnt this partially handled by InternetArchiveBot which is already active as global bot? If not, I think it would be easy to add this functionality. Myglobal bot converts a few http urls to https. This function of KiranBOT is a side-effect though — while converting the AMP urls to their canonical version, it is not the primary function. While this function could be integrated in KiranBOT, or InternetArchiveBot, a separate bot is preferred. Just a curious question. —usernamekiran(talk)03:35, 14 November 2025 (UTC)[reply]
      @Usernamekiran. Per my understanding, InternetArchiveBot identifies and replacesbroken external links. The link to the archive is then, of course, to a https site. This bot identifies and replaceslive external links that have an equivalent https site. On the 6 wiki sites this bot is already approved and since this request has started, the bot has performed another 90'000+ edits, showing that this particular task is not handled by other local or global bots. --Schlurcher (talk)10:44, 14 November 2025 (UTC)[reply]

    Removal of global bot status

    [edit]

    Please see theremoval policy for global bots for the requirements that must be met for global bot removal requests.

    Bot status requests

    [edit]

    Removal of bot status

    [edit]

    Requests for removal of the bot status from your own bots can be done here. If a specific user is abusing their bot status, you may request removal of the status on their account here as well. Please be sure to provide a short explanation on the current situation, along with links to the bot's contribution and an announcement in your community of the abuse. Previous requests arearchived.

    See also

    [edit]

    General requests for:help from a Meta sysop or bureaucrat ·deletion (speedy deletions:local·multilingual) ·URL blacklisting ·new languages ·interwiki map

    Personal requests for:username changes ·permissions (global) ·bot status ·adminship on Meta ·CheckUser information (local) ·local administrator help

    Cooperation requests for:comments (local) (global) ·translation

    Retrieved from "https://meta.wikimedia.org/w/index.php?title=Steward_requests/Bot_status&oldid=29679625"
    Categories:

    [8]ページ先頭

    ©2009-2025 Movatter.jp