Movatterモバイル変換


[0]ホーム

URL:


Jump to content
Wikimedia Meta-Wiki
Search

Talk:Spam blacklist

From Meta, a Wikimedia project coordination wiki
Latest comment:just now by Beetstra in topicDiscussion
Spam blacklist
Theassociated page is used by theMediaWikiSpam Blacklist extension, and listsregular expressions which cannot be used in URLs in any page in Wikimedia Foundation projects (as well as many external wikis). AnyMeta administrator can edit the spam blacklist; either manually or withSBHandler. For more information on what the spam blacklist is for, and the processes used here, please seeSpam blacklist/About.

Proposed additions
Please provide evidence of spamming on several wikis.Spam that only affects a single project should go to that project's local blacklist. Exceptions include malicious domains and URL redirector/shortener services. Please followthis format. Reports can also be submitted throughSBRequest.js. Please check back after submitting your report, there could be questions regarding your request.
Proposed removals
Please check ourlist of requests which repeatedly get declined. Typically, we do not remove domains from the spam blacklist in response to site-owners' requests. Instead, we de-blacklist sites when trusted, high-volume editors request the use of blacklisted links because of their value in support of ourprojects. Please consider whether requesting whitelisting on a specific wiki for a specific use is more appropriate - that isvery often the case.
Other discussion
Troubleshooting and problems - If there is an error in the blacklist (i.e. aregex error) which is causing problems, please raise the issue here.
Discussion - Meta-discussion concerning the operation of the blacklist and related pages, and communication among the spam blacklist team.
#wikimedia-external-linksconnect - Real-timeIRC chat for co-ordination of activities related to maintenance of the blacklist.
Whitelists
There is no global whitelist, so if you are seeking a whitelisting of a url at a wiki then please address such matters via use of the respectiveMediawiki talk:Spam-whitelist page at that wiki, and you should consider the use of the template {{edit protected}} or its local equivalent to get attention to your edit.

Pleasesign your posts with ~~~~ after your comment. This leaves a signature and timestamp so conversations are easier to follow.


Completed requests are marked as{{added}}/{{removed}} or{{declined}}, and are generallyarchived quickly. Additions and removals arelogged ·current log 2025/11.

Translate this page
Projects
Information
List of all projects
Overviews
Reports
Wikimedia Embassy
Project portals
Country portals
Tools
Spam blacklist
Title blacklist
Email blacklist
Rename blacklist
Closure of wikis
Interwiki map
Requests
Permissions
Bot flags
New languages
New projects
Username changes
Translations
Speedy deletions
[edit]
snippet forlogging
{{sbl-log|29718961#{{subst:anchorencode:SectionNameHere}}}}
SpBot archives all sections tagged with{{Section resolved|1=~~~~}} after 3 days and sections whose most recent comment is older than 7 days.

Proposed additions

[edit]
This section is for proposing that a website be blacklisted; add new entries at thebottom of the section, using the basic URL so that there is no link (example.com, nothttp://www.example.com).Provide links demonstrating widespread spamming by multiple users on multiple wikis. Completed requests will be marked as {{added}} or {{declined}} and archived.

apkpure.net

[edit]

apkpure.net

(LinkSearch:meta |en |es |de |fr |ru |zh |simple |c |d | Wikipedias: top25 ·50 ·major wikis ·sc ·gs)(Search:Google |en (G) |fr (G) |de (G) |meta (G) |backlinks |→ links ←)apkpure.net
(Reports:Report← track |XWiki |Local |en |find entry |blacklist hits)(DomainTools:whois |AboutUs |Malware?)

Dodgy app storespammed across many projects.Ohnoitsjamie (talk)13:58, 20 November 2025 (UTC)Reply

@Ohnoitsjamie:Added Added toSpam blacklist. --–DreamRimmer14:04, 20 November 2025 (UTC)Reply

apkpure.com

(LinkSearch:meta |en |es |de |fr |ru |zh |simple |c |d | Wikipedias: top25 ·50 ·major wikis ·sc ·gs)(Search:Google |en (G) |fr (G) |de (G) |meta (G) |backlinks |→ links ←)apkpure.com
(Reports:Report← track |XWiki |Local |en |find entry |blacklist hits)(DomainTools:whois |AboutUs |Malware?)

@DreamRimmer: One more: spammers have already switched to using .com variant.Ohnoitsjamie (talk)00:15, 26 November 2025 (UTC)Reply
@Ohnoitsjamie:Added Added toSpam blacklist. --–DreamRimmer02:06, 26 November 2025 (UTC)Reply

coupon-market.co.kr

[edit]

coupon-market.co.kr

(LinkSearch:meta |en |es |de |fr |ru |zh |simple |c |d | Wikipedias: top25 ·50 ·major wikis ·sc ·gs)(Search:Google |en (G) |fr (G) |de (G) |meta (G) |backlinks |→ links ←)coupon-market.co.kr
(Reports:Report← track |XWiki |Local |en |find entry |blacklist hits)(DomainTools:whois |AboutUs |Malware?)

Cross-wiki spammed to en, ja, ko, voyage, and wikidata.Ohnoitsjamie (talk)19:36, 21 November 2025 (UTC)Reply

@Ohnoitsjamie:Added Added toSpam blacklist. --🪶-TΛNBIRUZZΛMΛN (💬)19:43, 21 November 2025 (UTC)Reply

evtlindia.com

[edit]

evtlindia.com

(LinkSearch:meta |en |es |de |fr |ru |zh |simple |c |d | Wikipedias: top25 ·50 ·major wikis ·sc ·gs)(Search:Google |en (G) |fr (G) |de (G) |meta (G) |backlinks |→ links ←)evtlindia.com
(Reports:Report← track |XWiki |Local |en |find entry |blacklist hits)(DomainTools:whois |AboutUs |Malware?)

Spamming on multiple wikis with multiple accounts. --TonySt (talk)14:40, 25 November 2025 (UTC)Reply

@TonySt:Added Added toSpam blacklist. --Ternera (talk)15:26, 25 November 2025 (UTC)Reply

bl-india.com

[edit]

bl-india.com

(LinkSearch:meta |en |es |de |fr |ru |zh |simple |c |d | Wikipedias: top25 ·50 ·major wikis ·sc ·gs)(Search:Google |en (G) |fr (G) |de (G) |meta (G) |backlinks |→ links ←)bl-india.com
(Reports:Report← track |XWiki |Local |en |find entry |blacklist hits)(DomainTools:whois |AboutUs |Malware?)

Spamming on multiple wikis with multiple accounts. --TonySt (talk)14:41, 25 November 2025 (UTC)Reply

@TonySt:Added Added toSpam blacklist. --Ternera (talk)15:25, 25 November 2025 (UTC)Reply

etchejamz.com.ng

[edit]

etchejamz.com.ng

(LinkSearch:meta |en |es |de |fr |ru |zh |simple |c |d | Wikipedias: top25 ·50 ·major wikis ·sc ·gs)(Search:Google |en (G) |fr (G) |de (G) |meta (G) |backlinks |→ links ←)etchejamz.com.ng
(Reports:Report← track |XWiki |Local |en |find entry |blacklist hits)(DomainTools:whois |AboutUs |Malware?)

New link, same spam ring as in29651494.--A09|(pogovor)22:30, 26 November 2025 (UTC)Reply

@A09:Added Added toSpam blacklist. --A09|(pogovor)22:31, 26 November 2025 (UTC)Reply

easeus.com

[edit]

easeus.com

(LinkSearch:meta |en |es |de |fr |ru |zh |simple |c |d | Wikipedias: top25 ·50 ·major wikis ·sc ·gs)(Search:Google |en (G) |fr (G) |de (G) |meta (G) |backlinks |→ links ←)easeus.com
(Reports:Report← track |XWiki |Local |en |find entry |blacklist hits)(DomainTools:whois |AboutUs |Malware?)

They have been on a journey that is now 17 years long, and inbelow thread we still see that COI editors come here to request removal/whitelisting (and then they went on a spamming spree on en.wikipedia). Wikipedia is NOT for promotion of YOUR business. --Dirk BeetstraTC (en:U,T)10:56, 27 November 2025 (UTC)Reply

@Beetstra:Added Added toSpam blacklist. --Dirk BeetstraTC (en:U,T)10:56, 27 November 2025 (UTC)Reply

Proposed removals

[edit]
This section is for proposing that a website beunlisted;please add new entries at thebottom of the section. Use a suitable 3rd level heading and display the domain name as per this example{{LinkSummary|targetdomain.com}}. Please do not add the protocol part of domain name, eg.http

Remember to provide the specific domain blacklisted, links to the articles they are used in or useful to, and arguments in favour of unlisting. Completed requests will be marked as{{removed}} or{{declined}} and archived.

See alsorecurring requests for repeatedly proposed (and refused) removals.

Notes:

  • The addition or removal of a domain from the blacklist isnot a vote; please do not bold the first words in statements.
  • This page is for the removal of domains from the global blacklist, not for removal of domains from the blacklists of individual wikis. For those requests please take your discussion to the pertinent wiki, where such requests would be made atMediawiki talk:Spam-blacklist at that wiki.Search spamlists— remember to enter any relevant language code

Troubleshooting and problems

[edit]
This section is for comments related to problems with the blacklist (such as incorrect syntax or entries not being blocked), or problems saving a page because of a blacklisted link. This isnot the section to request that an entry be unlisted (seeProposed removals above).

Discussion

[edit]
This section is for discussion of Spam blacklist issues among other users.

Tooling / cleaning

[edit]

False signature to avoid archiving:Dirk BeetstraTC (en:U,T)00:00, 1 January 2026 (UTC)Reply

In#hometown.aol.co.uk we mentioned several ideas.

  1. It would be nice to have a tool that shows when an entry (or at least a specific domain or page) was last triggering a blacklist entry.
  2. As in 2015[1] (see alsoArchives/2015-01) we should delete old entries that have not triggered the SBL for x years. (x = 5?)
  3. It might be reasonable to move simple SBL entries, i.e. domains, (that are not locally whitelisted) to the global BED (list of blocked external domains). However,Special:BlockedExternalDomains is disabled. So it this an option now anyway?

Concerning 1.: I'm using a script for this. But for every domain it needs ~1000 db requests (one for each wiki). So I'm not sure, whether I should put that in a public web interface.--seth (talk)14:58, 5 October 2025 (UTC)Reply

Re 1. The URL for SBL hits is encoded in logging.log_params in a non-indexable way (see e.g.quarry:query/97741). To make that feasible we would need to collect hits in a user db. I have been thinking about doing this for spamcheck for quite a while.
Re 2. We could, but I don't see why that should be a priority IMHO.
Re 3. There is no global BlockedExternalDomains, seephab:T401524. Once this is implemented with a way to allow local whitelisting we can move stuff over.Count Count (talk)15:28, 5 October 2025 (UTC)Reply
1. Yes, using the `logging` table in the db is what I do in my script and what I also did in 2015. I'm using the replica db at toolforge. Using the replica directly, i.e. without having a separate db that contains the needed information only, searching all ~1000 wmf wikis takes about 1 or 2 minutes for a given regexp.
2. I mentioned reasons in the thread above. In short: performance. However,you don't need to do anything. I'd do that.
3. "Once it is implemented [...]": I see. So let's skip that for now.
--seth (talk)17:56, 5 October 2025 (UTC)Reply
1. I've backparsed the db once (and some of that data is in the offline linkwatcher database), however that is taking a lot of time, and since it are one-off runs it does not go up-to-date. Search engine would be nice per-wiki (just looking back for the lastn additions, with n defaulting to 2 or 3, looking backward for a choice timeframe), and one for cross wiki (with additional limitation for 'the big 5 wikis', 'the big 18 wikis + commons and wikidata', . For the application I suggested it does not have to find all additions, just the last couple.
2. I agree with the sentiment that it does not have priority, that performance loss is minimal, and I don't feel particularly worried if I blacklist 100 domains in one go that I bring the wiki down. Cleanup is good, though, it has a couple of advantages in administration as well (the occasional 'this website was spammed 15 years ago, it has now been usurped by another company', editing speed on the lists, easier to find complex rules).
3. BED really needs the whitelist to work on it, otherwiseespecially a global BED is going to be a pain for local wikis.Dirk BeetstraTC (en:U,T)06:02, 6 October 2025 (UTC)Reply
Unfortunately it seems that BlockedExternalDomain hit log entries are not being replicated to the Toolforge replicas. The log entries are just missing there. @Ladsgroup Is that on purpose?Count Count (talk)07:56, 6 October 2025 (UTC)Reply
Compare e.g.de:Special:Redirect/logid/140172685 andquarry:/query/97766Count Count (talk)07:59, 6 October 2025 (UTC)Reply
@Count Count Hi. I don't think that's on purpose and fixing it is rather easy. Would you mind creating a phabricator ticket assigning it to me with link to this comment? ThanksAmir (talk)15:32, 6 October 2025 (UTC)Reply
@Ladsgroup: Done, seephab:T406562. Thanks for having a look!Count Count (talk)10:19, 7 October 2025 (UTC)Reply
Thanks!Amir (talk)10:27, 7 October 2025 (UTC)Reply
1. I wrote a script to fetch all SBL data from all wmf wikis since 2020 and write all the data into a sqlite-db (the script needs ~7 minutes). This is not soo big (3,4M datasets in a 0,7GB db-file) and could be a) updated automatically (e.g. every day or every hour) and b) used in a little web interface to search the data. If I automatically delete all data that is older than 5 years, this might even scale. After the bug Count Count mentioned will be fixed, I could add the BED logs.
--seth (talk)22:00, 7 October 2025 (UTC)Reply
@Lustiger seth: Very cool. With this relatively miniscule amount of data I don't think that there even is a need to delete older data at all. It would be great if you could make the data available on a publicToolsDB database.Count Count (talk)04:28, 8 October 2025 (UTC)Reply
With that size I would even suggest to go back to the beginning of time. We do have some really long-term spamming cases (10+ or 15+ years). I ran into spamming of a website that related to one of those cases just last month. Having access to that (preferably through a query link in our {{LinkSummary}} (and maybe also the user templates) would be great.Dirk BeetstraTC (en:U,T)06:35, 8 October 2025 (UTC)Reply
Ok,the script created a databases51449__sbllog_p with one single tablesbl_log containing all sbl log entries of all wmf projects now and is continuously updated every 5 minutes. Its size is around 1,7GB and is has 8,1M entries. The columns are:id,project (e.g. 'dewiki'),log_id (local log_id),log_timestamp,log_namespace,log_title,comment_text,log_params (just the url),actor_name.
Next step is the creation of a web interface for queries. I'll try to do that on weekend.
--seth (talk)22:19, 9 October 2025 (UTC)Reply
Please both by user and domain so they can be linked from our {{LinkSummary}} and {{UserSummary}} templates.Dirk BeetstraTC (en:U,T)12:36, 10 October 2025 (UTC)Reply
@Lustiger seth: For faster domain querying Mediawiki (and spamcheck) store and index hostnames in reverse split order (e.g.www.google.com becomescom.google.www.. Maybe you could add such an indexed column and either keep the full URL or break it up in protocol + hostname + rest?Count Count (talk)13:30, 10 October 2025 (UTC)Reply
Because of RL things I haven't continued the work. I'll try to do something this weekend.
--seth (talk)09:32, 18 October 2025 (UTC)Reply
You can test it via:
It's very slow if date_from < 2025 or if the URL field does not contain anything with an extractable domain.
--seth (talk)14:12, 20 October 2025 (UTC)Reply
Oh, I'll remove the debugging output later, of course. I thought it might be helpful at the moment.
--seth (talk)14:14, 20 October 2025 (UTC)Reply
I have built it into the relevant templates here on meta and on en.wikipedia. Thanks!Dirk BeetstraTC (en:U,T)10:05, 21 October 2025 (UTC)Reply
@Lustiger seth Thanks!! Looks likedomain_rev_index is missing an index to make it fast though:quarry:query/98322Count Count (talk)10:35, 21 October 2025 (UTC)Reply
I did aCREATE INDEX idx_rev_index ON domain(domain_rev_index); now. Is it significantly faster now?
--seth (talk)13:28, 21 October 2025 (UTC)Reply
Oh yes. Just tried iton idealrentacar.ro and get results in less than a second, same for all other domains I tried it on.Count Count (talk)13:33, 21 October 2025 (UTC)Reply
Ah, ok, maybe my test was to early, because I still had to wait several dozens of seconds.
But now the results come fast for me, too.
--seth (talk)14:19, 21 October 2025 (UTC)Reply
Should I add an index to sbl_log.actor_name, too?
--seth (talk)14:54, 21 October 2025 (UTC)Reply
I think, that is a good idea. For spamcheck I am getting the global user id and storing that instead which survives renames and takes up a little less space but that is not really necessary IMHO.Count Count (talk)06:44, 22 October 2025 (UTC)Reply
I added an index for actor_name, now.
Global user id: I see, yes, makes sense. Let's hope that it's really not necessary. :-)
--seth (talk)20:49, 22 October 2025 (UTC)Reply
The query should probably be forWHERE d.domain_rev_index LIKE ? with the param being e.g.'com.chasedream.%' so we get hits forwww.chasedream.com as well. And each reversed hostname/domain should end with a '.' so that we can thus match both 'www.chasedream.com' and 'chasedream.com' but not 'chasedreamxyz.com' like this.Count Count (talk)13:42, 21 October 2025 (UTC)Reply
You are totally right. Should be done now.
--seth (talk)14:53, 21 October 2025 (UTC)Reply
Works great, thank you!Count Count (talk)06:40, 22 October 2025 (UTC)Reply
Ok, first step done. Maybe this or next weekend I'll have a look at the second step.
--seth (talk)08:49, 30 October 2025 (UTC)Reply
Hmm, haven't started yet. This needs more time.
Nevertheless, I found a bug in my scripts. The new database of the logs has several wrong entries, because some of the original entries in the wmf tables are really strange. I'll fix that bug first and then rebuild the tables.
--seth (talk)13:13, 9 November 2025 (UTC)Reply

Please add recorder.easeus.com to the spam whitelist

[edit]

recorder.easeus.com

(LinkSearch:meta |en |es |de |fr |ru |zh |simple |c |d | Wikipedias: top25 ·50 ·major wikis ·sc ·gs)(Search:Google |en (G) |fr (G) |de (G) |meta (G) |backlinks |→ links ←)recorder.easeus.com
(Reports:Report← track |XWiki |Local |en |find entry |blacklist hits)(DomainTools:whois |AboutUs |Malware?)


easeus.com

(LinkSearch:meta |en |es |de |fr |ru |zh |simple |c |d | Wikipedias: top25 ·50 ·major wikis ·sc ·gs)(Search:Google |en (G) |fr (G) |de (G) |meta (G) |backlinks |→ links ←)easeus.com
(Reports:Report← track |XWiki |Local |en |find entry |blacklist hits)(DomainTools:whois |AboutUs |Malware?)

Why: This page is the product homepage for EaseUS RecExperts software.Target pagesen:Comparison of screencasting software en:Comparison of webcam softwareUseAdd table row “EaseUS RecExperts” with neutral specs; no promo text.RemedyAffiliate spam removed; only documentation/release notes remain on subdomain.COII am an EaseUS employee, disclosed on talk pages.Thank you for your consideration.—{User:Cheesecreamalin} {14:50, 26 November 2025}Cheesecreamalin (talk)07:07, 26 November 2025 (UTC)Reply

The domain islocally blacklisted, you need to requestthere instead. 🪶-TΛNBIRUZZΛMΛN (💬)07:21, 26 November 2025 (UTC)Reply
I wouldn't bother, 17 years worth of spammy / non-notable additions. --Dirk BeetstraTC (en:U,T)10:51, 26 November 2025 (UTC)Reply
In fact, it should probably be added to the global blacklist after looking at spamcheck.Ternera (talk)14:05, 26 November 2025 (UTC)Reply
Done, who was saying that spammers 'surely move on after 8 years'..Dirk BeetstraTC (en:U,T)10:58, 27 November 2025 (UTC)Reply
Retrieved from "https://meta.wikimedia.org/w/index.php?title=Talk:Spam_blacklist&oldid=29718961"
Categories:
Hidden category:

[8]ページ先頭

©2009-2025 Movatter.jp