User:LiWa3/Settings: Difference between revisions
Appearance
From Meta, a Wikimedia project coordination wiki
adapt
that should not be here ... but now the other one \_OO_/
Line 282:
Line 282:
#domains for which we do not store webbugs and similar (regex on sortable domainname, i.e. the order of each part between the dots has been reverted, split multiple by |).
#domains for which we do not store webbugs and similar (regex on sortable domainname, i.e. the order of each part between the dots has been reverted, split multiple by |).
notracingstorage=archive\.org|web\.archive\.org
notracingstorage=archive\.org|web\.archive\.org
#Waiting times (in seconds) for re-loading lists of domains in xwiki and local categories for regenerating. Upon refreshing it records the time, and will not refresh the list until the time is passed. If refreshing all open reports takes longer than the set time, it will refresh continuously, otherwise it will wait until time is passed.
# local list waiting time set to 1 week.
locallistwait=604800
# xwiki list waiting time set to 12 hours.
xwikilistwait=43200
#</pre>
Revision as of 16:18, 18 January 2025
#
# Settings for LiWa3 # # Syntax is as follows: # * EVERY line starting with a "#" character is disregarded # * Every non-blank line is in the format "Setting=value" # # If you add channels to the <rcchannels> (separate channels with a space), # then you have to command 'LiWa3 restart diffreader' to make the bot listen to the new channels. # If you change on of the settings below, say '!reset' to implement settings in the bot (certain parts of the bot will have some lag) ircreportchannel=#cvn-sw-spam ircbotchannel=#wikipedia-spam-bot-channel ircalertchannel=#cvn-sw-spam ircelchannel=#wikimedia-external-links ircelchannels=#cvn-wp-es|#BeetstraBotChannel|#wikipedia-en-spam|#wikipedia-spam-t|#wikidata-spam|#wikimedia-external-links ownchannel=es.wikipedia.org|en.wikipedia.org|www.wikidata.org spam.en.wikipedia.org=#wikipedia-en-spam alert.en.wikipedia.org=#wikipedia-spam-t warning.en.wikipedia.org=#wikipedia-spam-t warn.en.wikipedia.org=#wikipedia-spam-bot-channel #spam.fa.wikipedia.org=#cvn-wp-fa #alert.fa.wikipedia.org=#cvn-wp-fa #warning.fa.wikipedia.org=#cvn-wp-fa warn.commons.wikimedia.org=#wikipedia-spam-bot-channel #alert.es.wikipedia.org=#cvn-wp-es #warning.es.wikipedia.org=#cvn-wp-es #spam.de.wikipedia.org=#cvn-wp-de #alert.de.wikipedia.org=#cvn-wp-de #warning.de.wikipedia.org=#cvn-wp-de spam.www.wikidata.org=#wikidata-spam alert.www.wikidata.org=#wikidata-spam warning.www.wikidata.org=#wikidata-spam # these are the wikis which have a XLinkBot blacklist, separate with | warningwikis=en.wikipedia.org|commons.wikimedia.org # This is where the local blacklists are on the wiki, format of parameter name is '<wiki>_blacklist', more lists can be # specified, separate with | en.wikipedia.org_blacklist=User:XLinkBot/RevertList|User:XLinkBot/RevertReferencesList|User:COIBot/RevertList|User:MER-C/RevertList.css en.wikipedia.org_overridelist=User:XLinkBot/OverrideList|User:MER-C/OverrideList.css|User:COIBot/OverrideList en.wikipedia.org_hardoverridelist=User:XLinkBot/HardOverrideList|User:MER-C/HardOverrideList.css|User:COIBot/HardOverrideList trustedgroups=sysop|bot|bureaucrat|rollbacker|founder|steward|checkuser|editor watchedspaces=0|2|6|10|14|108|118 # see http://en.wikipedia.org/w/api.php?action=sitematrix # The bot will read all that are mentioned behind 'site url="http://' # The selection of channels from 'special' needs to be specified here. separate with space special=#commons.wikimedia #meta.wikimedia #species.wikimedia #strategy.wikimedia #mediawiki.wikipedia #wikidata.wikipedia #outreach.wikipedia #nyc.wikimedia # reporter sleep time between posting a next report. # this number has to be tweaked according to the number of reporters active, # and the number of link-addition edits per minute (type: "LiWa3_1 !info" on IRC). # Example, 20 reported link additions per minute with 5 Reporters on IRC would each give 4 reports per minute # or one every 15 seconds. The sleeptime should then be smaller than 15 seconds (so it has also some time to work). # If it builds up a Queue (number after 'M'), either this time is too long, or one needs an extra reporter. # Setting this time too short may result in flooding. linkreportersleeptime=1 # linkwarnersleeptime=1 # linkparsersleeptime=1 # linkanalysersleeptime=1 # Using the parsed wikitext (parseddiff = 1) or the original unparsed diff (parseddiff = 0). useparseddiff=1 maxmessages=5 onelinkperline=1 # Do not save reports for every redirect detected. dontsaveredirects=0 #some wikis are too active and cause crashing of the diffreader. Therefore, they get their own. Crashing them also results then in downtime on these wikis. #Separate with space owndiffreader=#en.wikipedia #wikidata.wikipedia #some wikis are (extremely) high volume but are not having XLinkBot. That results in these still getting high priority and slowing down warning to XLinkBot. #wikis here will be pushed to low priority queue (queue3). Separate with | lowprioritywikis=#wikidata.wikipedia redirectsites=tk|trcx|attf tk_rule=\.tk\b tk_contentrule=\<frame\ssrc=\"(.*?)\" tk_ignorerule=\bdomain\.dot\.tk\b trcx_rule=\.tr\.cx\b trcx_contentrule=\<frame name=\"redirectframe\" src=\"((?:https?|ftp|irc|gopher|telnet|nntp|worldwind):\/\/.*?)\" attf_rule=\.at\.tf\b attf_contentrule=\bdocument\.location\s+=\s+\"(https?.*?)\"; checklocalwikis=1 checklocalgroups=0 largenumberoflinks=5000 largenumberoflinkadditions=2500 # if there are speed troubles, set this to 1 and the bot will report less stats and no XWiki/Local reports (but a significant # increase in speed) noextendedstats=0 noextendedstatslimit=500 lowstatlimit=250 # detect webbugs, adsenses and clickbanks and similar detectids=1 # if there are speed troubles, set this to 1 and the bot will report less stats and no XWiki/Local reports (but a significant # increase in speed) # This setting does the same as the above one (noextendedstats) but then ONLY for pages outside of mainspace. onlymainspacestats=0 onlymainspacestatslimit=2000 lowmainstatlimit=250 # dontstatwhitelistedusers=1 # dontscanwhitelistedlinks=1 # dontscanwhitelistedusers=1 # usercountmax=3000 usercountmax2=5000 autowhitelistusers=1 # linkcountmax=2500 numtemplates=4 template1=(?:https?|ftp|irc|gopher|telnet|nntp|worldwind):\/\/[^\s\]\[\{\}\\\|^`<>]+ template2=(?<![\w\d-])(?<!isbn[\s])\d{3,5}[\s-]\d{2,3}[\s-]\d{4}(?![\d\w-]) template3=(?<![\w\d-])(?<!isbn[\s])\d{6}[\s-]\d{4}(?![\d\w-]) template4=(?<![^\s:])[^\s\]\[\{\}\\\|^\/`<>@:]+@\w+(?!\.htm)(?:\.\w+){1,3} numalerts=2 alert1=\bproof\.php\b alert2=(?<![^\s:])[^\s\]\[\{\}\\\|^\/`<>@:]+@\w+(?!\.htm)(?:\.\w+){1,3} #alert3=(?<![\w\d-])(?<!isbn[\s])(?<!issn[\s\%\(])(?<!\%)\d{3,5}[\s-]\d{2,3}[\s-]\d{4}(?![\d\w-]) #alert4=(?<![\w\d-])(?<!isbn[\s])(?<!issn[\s\%\)])(?<!%)\d{6}[\s-]\d{4}(?![\d\w-]) # Throw an alert when editors add more than XX external links in ONE edit externallinkoverflow=100 numignores=7 # ignore base wiki links. ignore1=https?:\/\/(?:wiki(?:[mp]edia|books|species|source|versity|news)|wiktionary|mediawiki)\.org\/[\?w] # ignore email addresses from mediawiki projects ignore7=@[A-Za-z\-]+\.(?:wiki(?:[mp]edia|books|species|source|versity|news)|wiktionary)(?!\.) # ignore the toolserver ignore2=tools\.wikimedia\.de\b # ignore mediawiki projects themselves ignore3=https?:\/\/[A-Za-z\-]+\.(?:wiki(?:[mp]edia|books|species|source|versity|news)|wiktionary|mediawiki)\.org\/[\?w] # ignore more toolserver ignore4=\btoolserver\.org\b # ignore links that are not links ignore5=^http:\/\/\',$ # ignore some malformed links ignore6=^http\.$ # regex for things that do not need to be scanned for adsense/clickbank and such noidsensing=\.pdf\b|\.doc\b|\.jpg\b|\.gif\b #For the stopforumspam.com queries - reporting when any of these is passing: #confidence threshold - only report when 'confidence' is higher than this (value between 0-100) # Lets try 25% ## bump to 50% as the username is least reliable, and probably IP address more of a concern forumspam_confidencethreshold=50 #frequency threshold - only report when 'frequency' is higher than this # Lets try minimum of 10 forumspam_frequencythreshold=10 #time threshold - only report when editor was last seen on stopforumspam.com within the last ## seconds - set to 0 for infinite. # Lets try last 4 weeks for users, 1 week for IPs forumspam_usertimethreshold=2419200 forumspam_iptimethreshold=604800 #username length threshold - only report when the length of the non-IP username is longer than this # Lets try minimum username length of 3, single and double character usernames are too common. forumspam_usernamelengththreshold=2 # Some on-IRC tags - MAKE SURE TAGS ARE UNIQUE, THEY ARE USED IN THE CODE TO TEST AS WELL tag_onlymainspacestats=MSO tag_noextendedstats=NES tag_notcounted=NC tag_dontcount=DC tag_clickbanklink=CB tag_webbuglink=WB tag_adsenselink=AS tag_plimuslink=PL tag_autoredlist=ARL tag_redlist=RL tag_alertlist=AL tag_redirectlink=REDIR tag_whitelistedlink=WL tag_blacklistedlink=BL tag_resolvedip=RESOLVED tag_forumspammer=FS tag_status1XX=STATUS CONTINUE tag_status100=STATUS 100 OK tag_status101=STATUS 101 Continue tag_status2XX=STATUS OK tag_status200=STATUS 200 (OK) tag_status201=STATUS 201 (created) tag_status202=STATUS 202 (accepted) tag_status203=STATUS 203 (non-authorative information) tag_status204=STATUS 204 (no content) tag_status205=STATUS 205 (reset content) tag_status206=STATUS 206 (partial content) tag_status300=STATUS 300 (multiple choice) tag_status301=STATUS 301 (moved permanently) tag_status302=STATUS 302 (found) tag_status303=STATUS 303 (see other) tag_status304=STATUS 304 (not modified) tag_status305=STATUS 305 (use proxy) tag_status306=STATUS 306 (not used) tag_status307=STATUS 307 (temporary redirect) tag_status300=STATUS 400 (bad request) tag_status401=STATUS 401 (unauthorized) tag_status402=STATUS 402 (payment required) tag_status403=STATUS 403 (forbidden) tag_status404=STATUS 404 (not found) tag_status405=STATUS 405 (method not allowed) tag_status406=STATUS 406 (not acceptable) tag_status407=STATUS 407 (proxy authentication required) tag_status408=STATUS 408 (request timeout) tag_status409=STATUS 409 (conflict) tag_status410=STATUS 410 (gone) tag_status411=STATUS 411 (length required) tag_status412=STATUS 412 (precondition failed) tag_status413=STATUS 413 (request entity too long) tag_status414=STATUS 414 (request-URI too long) tag_status415=STATUS 415 (unsupported media type) tag_status416=STATUS 416 (requested range not satisfiable) tag_status417=STATUS 417 (expectation failed) tag_status500=STATUS 500 (internal server error) tag_status501=STATUS 501 (not implemented) tag_status502=STATUS 502 (bad gateway) tag_status503=STATUS 503 (service unavailable) tag_status504=STATUS 504 (gateway timeout) tag_status505=STATUS 505 (HTTP version not supported) tag_noredirectlink=NOREDIR # More tags - just for display, no need to be unique tag_veryhighcount=!! tag_highcount=! tag_highalert=!! tag_lowalert=! # Since we have some bots running around doing A LOT of work, making life for linkwatcher impossible: # totallyignoredusers, is a '|' separated list, either containing: # a username in each item (which is then globally ignored), or # an item 'username@wiki' (username and wiki separated with @) for ignoring only the user on a certain wiki (wikiname is as in the url, 'meta.wikimedia.org', 'www.wikidata.org', 'en.wikipedia.org') totallyignoredusers=Legobot|Addbot|Sk!dbot|COIBot|GZWDer (flood)@www.wikidata.org|S205643bot@www.wikidata.org|QuickStatementsBot@www.wikidata.org|BotNinja@www.wikidata.org|ProteinBoxBot@www.wikidata.org|JarBot@ar.wikipedia.org # totallyignoredgroups, is a '|' separated list containing an item 'right@wiki' (right and wiki separated with @) for ignoring all the with a certain right on a certain wiki (as just above for totallyignoredusers) # global right ignoring is NOT supported. totallyignoredgroups=flood@www.wikidata.org|bot@www.wikidata.org|bot@commons.wikimedia.org #This determines the places for the on-wiki whitelists and similar wikilinkwhitelist=User:LiWa3/linkwhitelist.css wikiuserwhitelist=User:LiWa3/userwhitelist.css wikidonotcountdomainlist=User:LiWa3/donotcountlist.css # These are cloaks of trusted users who can always use all commands of COIBot: trustedusers=user/beetstra|wikimedia/Versageek|wikipedia/MER-C|wikimedia/Igna|wikipedia/pdpc.professional.kaare|wikimedia/Martin-Urbanec|wikisource/billinghurst|wikipedia/Trijnstel|wikipedia/Shanmugamp7|wikimedia/Glaisher|wikimedia/marcoaurelio|wikimedia/-revi|wikimedia/Zabshk|countervandalism/Syum90|wikipedia/TheresNoTime|wikimedia/JonKolbert|wikipedia/Praxidicae|wikimedia/L235|wikipedia/ZI-Jony|wikipedia/GeneralNotability|wikimedia/ToBeFree|wikipedia/perryprog #domains for which we do not store webbugs and similar (regex on sortable domainname, i.e. the order of each part between the dots has been reverted, split multiple by |). notracingstorage=archive\.org|web\.archive\.org