This is an archive of past discussions. Do not edit the contents of this page. If you wish to start a new discussion or revive an old one, please do so on the current talk page. |
Archive 65 | ← | Archive 68 | Archive 69 | Archive 70 | Archive 71 | Archive 72 | → | Archive 75 |
Over the years, I have come into situations where someone will slap a merge tag on an article, and then not follow through with adding a merge tag to the other article, if they even decide to mention why they want to merge the two pages at all. These tags can often remain up for years until they are removed, so I was wondering if there was a way to program a bot to remove these sorts of things, as I suspect a sizable chunk of merge request taggings are just that. This may be near impossible to do, but it would be worth looking into, if possible. Kevin Rutherford ( talk) 17:33, 21 January 2016 (UTC)
How is a bot supposed to know that there's no discussion? What if a discussion is started with a heading like aren't these the same thing or do we really need two articles. A human can figure out that those are likely merge discussions, but a bot can't. Oiyarbepsy ( talk) 14:42, 18 February 2016 (UTC)
For some reason, loads of our articles cite The Minor Planet Bulletin as Bulletin of the Minor Planets Section of the Association of Lunar and Planetary Observers, which is sometimes used as a subtitle to The Minor Planet Bulletin (see example). The title of the publication itself is, however, The Minor Planet Bulletin, and that's how the journal should be cited. So if someone could do a simple search/replace for
on the following pages
That would be much appreciated. I started doing it, but with ~230 articles to go through, this is much more suited for a bot. Headbomb { talk / contribs / physics / books} 00:20, 2 May 2016 (UTC)
|journal=The Minor Planet Bulletin, Bulletin of the Minor Planets Section of the Association of Lunar and Planetary Observers
, as on
10000 Myriostos. Is it desired to truncate these to |journal=The Minor Planet Bulletin
or leave them as-is? FWIW, I prefer the shorter version, both aesthetically and procedurally as the actual title. ~
Tom.Reding (
talk ⋅
dgaf) 23:55, 2 May 2016 (UTC)
The following should have the following replacement made
|publisher=Minor Planets Section of the Association of Lunar and Planetary Observers
). I'll post the list of changes here when I'm done, so that all these related changes are in 1 thread, for future reference. ~
Tom.Reding (
talk ⋅
dgaf) 00:55, 4 May 2016 (UTC)As of now, I have also permanently switched the journal-parameter from "Bulletin of the Minor Planets" to "The Minor Planet Bulletin" and removed the obsolete publisher-parameter. I'll retroactively amend my edits back to May 1st. Thank you all for your efforts. Rfassbind – talk 07:24, 4 May 2016 (UTC)
What would everyone think of a bot that auto-cleaned up the comments left behind by new articles created with the article wizard? For reference, I'm referring to these. -- Nathan2055 talk - contribs 04:25, 6 February 2016 (UTC)
It would be useful to have a bot that patrolled articles on fringe theories, to stop any fringe links from being added and delete passages which give undue weight. This is because there is a problem with UFO articles at the moment, many are not written according to fringe guidelines. I need someone to help me with making this, I can't code, but I want to learn how to code bots. ThePlatypusofDoom (Talk) 17:07, 12 May 2016 (UTC)
Do we have a bot that could update Wikipedia:Database reports/Long pages, say, monthly? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 21:36, 20 March 2016 (UTC)
There are hundreds and hundreds (probably thousands) of articles about sport teams (expecially football) which are completely unreferenced, and, as most of articles about sport teams contains infos like establishment/disestablishment year, honours, staff people names etc., it's very important to have references in such type of articles. So, I propose to scan WP dump for articles about sport teams, without any external link in them, without <ref>
tags, and probably without {{
Reflist}} tag and/or "References"/"Notes" section - thus list of pages to go through is formed - and then to go through them and to tag with {{
unreferenced}}. When running to do the job also it's necessary to parse source code and visual output of articles to ensure that there are no refs before tagging page. --
XXN, 18:15, 2 March 2016 (UTC)
<tt>
elementThe <tt>...</tt>
element no longer even exists in HTML. Like <font>
, it was a
separation of style and content problem. While MediaWiki doesn't choke on it, we need to stop using it like it's valid, and replace it with something more appropriate.
<span style="font-family: monospace;">...</span>
. We can't be certain that code examples were the intended use, and in many cases they were not. For the cases that are, the span markup isn't the most perfect possible markup for such examples, but using the span tag isn't "wrong", where the code tag often would be, and if anyone wants to, they can replace the span later.<code>...</code>
, as the tt dead-element is frequently used incorrectly to mark up code examples in template documentation. In the odd case that the actual output of a template uses the tt tag and is not for code (is there a template for representing telegram output? I doubt it, but it's possible), the special appearance of the code output
compared to the span output will make it obvious that the template needs to be tweaked to use a monospaced span.In special namespaces like Module and MediaWiki, uses of it should simply be identified and listed, not altered. It can safely be replaced in mainspace, Wikipedia, Portal, all talk namespaces, etc. (except where it's in nowiki or source xtags).
The bot could generate a list of templates that do not have names ending in /doc in which this substitution was performed, so they can be checked manually to ensure they don't need their ouput changed to use <span>
(or <samp>
or <kbd>
). Or just listed and not changed, and slated for manual cleanup. I'll volunteer to do that part of the cleanup, either way. —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 18:58, 19 February 2016 (UTC) Clarified. —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 21:26, 1 March 2016 (UTC)
PS: I have no objection to it being added to AWB General Fixes instead, if people think that having humans do it incrementally would be better. As long as the cleanup begins one way or another. — SMcCandlish ☺ ☏ ¢ ≽ʌⱷ҅ᴥⱷʌ≼ 19:08, 19 February 2016 (UTC)
<ref name=foo/>
(with or without quotes around the value of name=
) and <br/>
and <br>
(and the worse error "</br>
") to use proper ... />
(i.e., space, then slash, then close-angle-bracket) syntax. Preferably it would quote the value of name=
(MediaWiki doesn't choke on it, but any number of external XML parsing tools will). Another obvious fix of this sort is putting quotes around any [X|H]TML attribute values that are non-numeric, e.g. fixing class=classname
to class="classname"
(it's actually safest to quote all of them, since a numeric one could be changed to non-numeric at any time, just like any name=Johnson1999
could be changed to name=Johson 1999
by a later editor (and even MW will barf on that). The same sort of cleanup script could also perform dead-code cleanup in the form of converting any empty <element [attribute=value [...]]></element>
to <element [attribute=value [...]] />
. That last one would be arguably cosmetic and thus better done as an AWB general fix. A different kind of tool might also build up a list of pages with the same HTML id=
value used two or more times on same page, for manual fixing. I've long wondered if whatever trick is used by navboxes to detect if another navbox is present and auto-collapse could also be used to catch this error, in templates that generate id
s. Would also be nice to track down uses of <font>...</font>
(which can be auto-converted to <span>...</span>
, with some work) and <center>...</center>
(which would require manual fixing). Oh, and <acronym>...</acronym>
should in every case be changed to <abbr>...</abbr>
; the former has been invalid for years, and they support the same attributes with the same output. —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 21:26, 1 March 2016 (UTC)
On breaks: <br /> is actually only a work around for some legacy HTML parsers and is not strictly correct, though an XML parser won't choke on it (for those authors who want to serve XHTML). I doubt anything in this day and age which we support requires it. More generally, <br> is fine in Html 5, which we've been serving for years at this point. ( See StackOverflow.) So I would not approve of any change but </br> -> <br(|/| /)>. This should probably be a general fix or a CHECKWIKI fix if it isn't already.
I would support a bot to add quotation marks to any attributes, per this recent change, soon-live onwiki.
The general <element></element> -> <element/> has some issues on some legacy browsers for some attributes, related mostly to the "/" (and especially in block elements e.g. divs and paragraphs). I think these are the same legacy browsers as with the break problem.
I think the Javascript looks for the collapsible class, so you'd need to check whether "getElementByID" returns a list or a single element. Probably the latter, since an ID is supposed to be unique.
I would support replacement of font, center, and acronym, though it may be better just to template-ize them for legitimate uses and remove them for illegitimate. Probably a better task for AWB, if not AWB general fixes. -- Izno ( talk) 21:56, 1 March 2016 (UTC)
<br />
is strictly correct; XML (including XHTML) requires the /
. HTML5 does not require it, but it works fine in HTML5, and people reuse our code in more ways that we think they do, so it is best to give them the most portable code. The <br/>
(no space) format, which is also valid XML, is what certain old browsers have a problem with. It is the exact same problem as <hr/>
or any <element/>
markup, without the space, in the same browsers. The <element></element>
→ <element />
conversion does not present any problems, only <element></element>
→ <element/>
conversion. —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 13:42, 6 March 2016 (UTC)<font>
Between a bot and alot of *$(#& manual fixes, there are currently no <font>
tags in article space.<big>
But, it is not currently switched on for enwiki.</br>
There are currently no </br>
tags in articles, but alot are added everyday. Checkwiki also checks for other cases of bad <br>
tags.<tt>
to Checkwiki. I can also give a listing of articles using it via a dump file or any other tags. At one point, I was changing <tt>
to {{
mono}} or <code>
, but was getting alot of complaints, so I dropped it.
Bgwhite (
talk) 22:53, 1 March 2016 (UTC)
<big>...</big>
and <small>...</small>
would be nice to get rid of. They're convenient for entry, like <font>...</font>
, but we should not be using these things in mainspace or any equivalent, including template, book, portal, etc. It would be nice if all this stuff were purged from Wikipedia namespace, too. Really, everywhere, though I guess we care least of all if they remain in the talk spaces. People can complain all they want, but <tt>...</tt>
in particular is just dead and they have to move on with their lives. That whole "acceptance" stage of grieving. :-) —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 13:42, 6 March 2016 (UTC)
<tt>
tags from the March dump can be found at
User:Bgwhite/Sandbox1. There are sports articles at the beginning of the list in which the tt tag should be removed and not converted to <code>
. Not all tt tags should be converted to code, so a bot couldn't run on the list. Manual editing should be done.
Bgwhite (
talk) 18:51, 7 March 2016 (UTC)
{{
mono}}
in 2008 or so, because <tt>...</tt>
was frequently being used in articles. A bot or AWB script or whatever could add a silent parameter for tracking purposes. —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 18:59, 26 March 2016 (UTC)@
SMcCandlish: As usual, you've completed the work before I've even thought of it. I'd recommend placing that template around the text currently in tt tags, with additional parameter | needs_review = yes
placed at the end (or beginning, whichever's easier). Someone can create a tracking category of every transclusion with a nonempty needs_review
parameter at a later time. To be clear, this is not a substitute for properly informing editors that the edit is worth reviewing via the edit summary. A proper implementation should do both. ~
Rob
Talk 19:34, 26 March 2016 (UTC)
\<tt\>([^\<]*)\<\/tt\> --> {{mono|$1|needs_review=yes}}
on whichever namespaces this task has consensus to run on. It would miss something that had additional HTML tags within the tt tags, but it should hit most things, and the rest probably are worth editor review anyway. Where is the consensus discussion for this, by the way? ~
Rob
Talk 00:25, 31 March 2016 (UTC)Be careful about your proposed changes to the name=
tags. The danger points are, offhand 1. The possible (though horrible) situation where Johnson1999 and "Johnson1999" are different references. and 2. Accidental breaking of citation templates, e.g. the Harvard Referencing templates.
As for the rest, one issue is that this makes the code somewhat harder for editors to understand. I'd suggest using templates wherever the code gets complicated. Complicated, naked HTML is an accessibility issue. . Adam Cuerden ( talk) 00:33, 31 March 2016 (UTC)
name=foo/
is also distinct from name=foo
then I think that only the accessibility issue (templating instead of complicated raw HTML) remains. Perhaps {{
mono}}, which already exists?
Adam Cuerden (
talk) 00:43, 31 March 2016 (UTC)
<strike>...</strike>
is <s>...</s>
. WP emits a doctype of plain "HTML" (i.e., HTML5), not HTML 4.0 strict, anyway. If the argument is that any strike-through, including <s>...</s>
and <del>...</del>
is an accessibility issue, that seems to be a more theoretical matter; both those tags are part of the HTML specs, and this thread is about complying with them but getting rid of use of the invalid <tt>...</tt>
element. —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 09:19, 1 April 2016 (UTC)Template:Filmr and Template talk:Filmr both show up at Special:WantedPage with about 3000 incoming links each. The template itself was deleted in 2012 since it was just a holding page. It's used in many non-free screenshots as part of the fair use rationale parameter within the Template:Navbox used for the rationale. Could a bot remove the entire name parameter from these Navbox templates on the pages that are linking to Filmr? The box will then stop trying to link to the deleted template, talk and edit option as seen here and instead will be a plain box like this. Either the parameter can be removed or a bot could just as easily remove the text "Filmr" to reduce the chance of error. -- Ricky81682 ( talk) 09:26, 30 March 2016 (UTC)
{{
Navbox}}
with a custom template. All the best:
Rich
Farmbrough, 00:17, 3 April 2016 (UTC).Could somebody recreate User:Pigsonthewing/Direct calls to Infobox for me, please? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:59, 25 April 2016 (UTC)
\{\{\s*([Ii]nfo box|[Ii]nfoBox|[Ii]nfobox [Cc]onditionals|[Ii]nfobox/old|[Ii]nfobox/row|[Rr]ow)\s*[\|\}]
, which includes all of the #Redirects to {{
Infobox}} (and only #Redirects), returned these 3 additional pages:
Windows NT 4.0,
Windows Embedded Automotive,
Windows Embedded Industry. ~
Tom.Reding (
talk ⋅
dgaf) 15:38, 4 May 2016 (UTC)
WP:MOSTEDITED is sorted by both the number of editors and the number of edits, based on a formula to convert them both into standard scores from their approximately lognormal distribution as non-negative integers, and then adding the resulting normal scores together before sorting to rank the articles.
I want to do the same thing with both pageviews and ORES wp10 quality scores. The pageviews API is at e.g. https://wikimedia.org/api/rest_v1/metrics/pageviews/ ( per-article documentation) and the ORES API is at https://ores.wmflabs.org/v2/#!/scoring/get_v2_scores_context_model_revid
I want to start with, say the top 1,000 articles from May 1, 2016 and get a list of those with the lowest ORES scores, and go from there. Who can help me please? EllenCT ( talk) 14:40, 10 May 2016 (UTC)
Done as per below; shortcut WP:POPULARLOWQUALITY. EllenCT ( talk) 05:08, 23 May 2016 (UTC)
Here's what I have so far for trying to find the lowest quality high-popularity articles:
code
|
---|
import sys import urllib2 import safeJSON r1 = urllib2.Request(' https://wikimedia.org/api/rest_v1/metrics/pageviews/top/en.wikipedia.org/all-access/2016/05/15') # use about 5 days ago (test how long ago is reliable) for the date at the end of that url in production e.g.: # import datetime # print ((date.today() - timedelta(days=5)).strftime("%Y/%m/%d")) try: u1 = urllib2.urlopen(r1) top_articles = safeJSON.loads(u1.read()) for i in range(2, 999): article = top_articles['items'][0]['articles'][i]['article'] views = top_articles['items'][0]['articles'][i]['views'] title = article.replace('_', ' ') r2 = urllib2.Request(' https://en.wikipedia.org/w/api.php?action=query&format=json&prop=revisions&titles=' + article) try: u2 = urllib2.urlopen(r2) revinfo = safeJSON.loads(u2.read()) revid = revinfo['query']['pages'][revinfo['query']['pages'].items()[0][0]]['revisions'][0]['revid'] r3 = urllib2.Request(' https://ores.wmflabs.org/v2/scores/enwiki/wp10/' + str(revid)) try: u3 = urllib2.urlopen(r3) ores = safeJSON.loads(u3.read()) prediction = ores['scores']['enwiki']['wp10']['scores'][str(revid)]['prediction'] if prediction in ['Stub', 'Start', 'C']: print('[' + '[%s]] %s-class with %d views' % (title, prediction, views)) except: 0 # print(i, title, 'no ORES score') # e.g. article is too new except: 0 # print(i, title, 'can not get latest revision') # e.g. Special: namespace except: print('can not load top articles') # e.g., less than 5 or so days ago |
Can someone who understands bots please turn that into a WP:MOSTEDITED-style report updating daily with the most recent top articles date (at the end of the first url)? EllenCT ( talk) 09:30, 22 May 2016 (UTC)
@ Bamyers99: can you please add this to DataflowBot and link the daily report from WP:BACKLOG? Ideally the output would be in similar tabular format sorted first by Stub, Start, and C-class prediction, then by descending pageviews. EllenCT ( talk) 18:13, 22 May 2016 (UTC)
@ Bamyers99: would you please make the warning added to the top of the report permanent until issues with pageview transience (I think we can use a simpler version of the algorithms described here) and ORES predictions for articles composed mostly of templates get sorted out? EllenCT ( talk) 15:15, 23 May 2016 (UTC)
@ Bamyers99: please see the suggestions here for some recommended improvements to the warning message. I am trying to get a top-100,000 list so we may be able to drop the C-class predictions, which aren't generally in urgent need of improvements. EllenCT ( talk) 12:57, 25 May 2016 (UTC)
Further discussion on the bot's talk page please. EllenCT ( talk) 21:15, 27 May 2016 (UTC)
Some time ago, the parameter |licensor=
was corrected to |licensee=
, however the template calls in article space were never updated to reflect the name change. Now that |licensor=
is being phased out, the call will need to be updated.
I would have done this myself using AWB, but considered it to trivial of an edit to be allowed under AWB usage rules. — Farix ( t | c) 11:38, 21 May 2016 (UTC)
@ TheFarix: now the change is part of AWB general fixes. It is easy to be done now. -- Magioladitis ( talk) 16:07, 23 May 2016 (UTC)
Hello all! Not sure if this is the correct place to request this, so let me know if I should go elsewhere. I am working on restarting the Wikipedia:WikiProject Green Bay Packers and would like to get an updated and completed list of all of the articles that have had some type of recognition (featured articles, featured lists, DYK, etc.). I would like a query to be run to generate a list of Green Bay Packers-related content that some form of recognition. This could be done by finding searching all pages that are in Category:WikiProject Green Bay Packers and querying out those that have {{ Article history}} and {{ DYK talk}}. I am also open to other ideas that would complete the task. The main output would be a list or relevant articles (which could be placed here). Let me know if this can be done, or if I should ask elsewhere. Appreciate the help! Thanks, « Gonzo fan2007 (talk) @ 01:18, 30 May 2016 (UTC)
More than a week ago I placed my #Bot request above, but nobody responses. What's going wrong? Sander.v.Ginkel ( Talk) 09:38, 25 February 2016 (UTC)
I can start the tagging after the weekend. -- Magioladitis ( talk) 10:13, 25 February 2016 (UTC)
I've suggested before that instead of simply archiving this page, requests should be archived in one of two ways: "done/won't do/not valid" vs. "awaiting action". That way, valid requests that are not recent will be more readily found by someone who might want to work on them. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:08, 14 March 2016 (UTC)
Most well-developed Wikipedia pages of a species have information from the IUCN Red List of Threatened Species. Unfortunately there doesn't seem to be an automated process to keep the information up to date.
At the moment the data on the page for Gorilla gorilla was added in 2009. Wikipedia thus lists incorrectly "Current Population Trend" as unknown when the IUCN list it as "decreasing". I imagine that there are a lot of pages where the information that's in Wikipedia is outdated. Many smaller pages of species such as the Unicorn leatherjacket completely lack the information, while IUCN lists the information on the status of Unicorn leatherjacket (Aluterus monoceros) on their page.
Given that the IUCN reports their data in a very orderly fashion it should be straightforward to write a bot that regularly transfers new data from ICUN to Wikipedia. — Preceding unsigned comment added by ChristianKl ( talk • contribs) 14:58, 3 April 2016 (UTC)
It's been a while since we've had a tagging run the project, so if a bot could tag the following articles with {{ WikiProject Academic Journals}} this would be great. Were' talking
|class=
from other banners when possible.
|class=
should be used.|needs-infobox=yes
. Some articles won't need an infobox, but they can be dealt with one by one.|needs-infobox=yes
that have an infobox should be untagged. They can be found at
Category:Journal articles needing infoboxesAdditionally
|cover=
of {{
Infobox journal}} should be tagged with {{
WikiProject Academic Journals}}.Headbomb { talk / contribs / physics / books} 15:59, 5 April 2016 (UTC)
Headbomb I can do it. -- Magioladitis ( talk) 16:07, 5 April 2016 (UTC)
Headbomb Please leave a message at Wikipedia talk:WikiProject Academic Journals to link to this discussion here so everyone is aware about it. Thanks, Magioladitis ( talk) 16:27, 5 April 2016 (UTC)
Headbomb I know but after all these complains I keep getting I want to be on the safe side. I also still look for volunteers to do the tagging instead of me. -- Magioladitis ( talk) 20:52, 5 April 2016 (UTC)
Headbomb For instance United Nations Economic Commission for Europe contains Infobox journal. -- Magioladitis ( talk) 21:38, 5 April 2016 (UTC)
I auto-assessed for Stub/Start class and manually fixed all in NA-class. -- Magioladitis ( talk) 10:28, 7 April 2016 (UTC)
Headbomb I can't find an easy to way to fulfill the images request. -- Magioladitis ( talk) 10:33, 7 April 2016 (UTC)
I am so happy to discover that I was the one who did a previous bot tagging in 2011. 5 years ago :) -- Magioladitis ( talk) 11:21, 7 April 2016 (UTC)
The subheaders of Wikipedia:Administrators' noticeboard/Requests for closure are "Requests for comment", "Backlogs", "XfD", "Administrative", and "Requested moves". For each subheader, add in the <includeonly> section that is visible only to WP:AN (to which WP:ANRFC is transcluded) how many discussions are waiting closure in that section and how old the oldest discussion in that section is. Maybe update this once a day. Pinging BU Rob13 ( talk · contribs), who suggested this here. Cunard ( talk) 20:46, 10 April 2016 (UTC)
I write in my capacity as one of two Wikipedians in Residence at TED.
Firstly, I would appreciate some help, please, in counting links to TED talks (URLs including [Now resolved.]
/talks/
), topics (URLs including /topics/
) and speaker profiles (URLs including /speakers/
) in this and the other top ten largest Wikipedias. There is some prior discussion of the issue at
WP:VPT#External links by page type.
Secondly, it would be a good idea to clean up TED links. We have both external links and links in citations, to pages like:
and:
In such links, the index.php/
and .html
parts are redundant; and http://
would be better replaced by https://
Links may have one, two, or all three of these issues.
Thirdly, external links for TED speakers should be replaced using {{ TED speaker}} - though that might be better done once Wikidata is populated with TED IDs, and values can be called from there. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 16:57, 7 March 2016 (UTC)
Fourthly, links like https://ted.com/daphne_koller_what_we_re_learning_from_online_education should have the "www" prefix added. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 13:54, 8 March 2016 (UTC)
{{ TED talk}} is now available, also. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:04, 9 March 2016 (UTC)
I've now determined that there are
148 links whose URL includes ted.com/index.php/
, and
883 links whose URL ends with .html
.
Andy Mabbett (Pigsonthewing);
Talk to Andy;
Andy's edits 11:26, 14 March 2016 (UTC)
@ Hasteur: Thank you. Yes, except:
https://
is
preferred.-- Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:03, 14 March 2016 (UTC)
Update: Links in the form http://www.ted.com/talks/lang/en/thandie_newton_embracing_otherness_embracing_myself.html
should also drop the lang/en/
component (for all langauge codes; 93 instances).
Andy Mabbett (Pigsonthewing);
Talk to Andy;
Andy's edits 20:36, 14 March 2016 (UTC)
As Hasteur is unavailable, can anyone else help with this, please? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 15:10, 14 April 2016 (UTC)
Wikipedia Zero's made us attractive as a piracy host ( T129845). To combat music piracy I suggest setting up an audio fingerprinting system like Echoprint or AcoustID. These are open sourced (unlike Shazam or Gracenote) and supposedly easy to get running. — Dispenser 23:27, 14 April 2016 (UTC)
Can someone take over for User:Wikinews Importer Bot since Misza13 seems to have disappeared about a year ago? It would be greatly appreciated. ··· 日本穣 · 投稿 · Talk to Nihonjoe · Join WP Japan! 01:27, 17 February 2016 (UTC)
I appreciate the update. I wish you luck on finding that kind soul. ··· 日本穣 · 投稿 · Talk to Nihonjoe · Join WP Japan! 16:00, 21 April 2016 (UTC)
Hi,
Can someone please sort the names in Category:Wives of Ottoman Sultans and Category:Daughters of Ottoman sultans so that they are arranged by first name and not by title? These people don't actually have a last name (see List of Ottoman titles and appellations and a corresponding category at Turkish wikipedia) and right now their categories are chaotic, some of them are arranged by first name and some of them by title. (I expect this to be a recurring problem as some editors might not be familiar with the fact that not everything is a surname that occurs at the end of a name...) I wanted to sort them out but there are over 100 of them.
Thanks, – Alensha talk 17:12, 28 May 2016 (UTC)
Currently there exists a log of all blocks, which can be examined. The fact that it exists and can be examined is crucially important to transparency. There exists no such log of unblock requests which links to them/lists whether or not they were granted. Therefore, there is no practical manner by which to get a big picture sense of what is going on in the realm of unblock requests; there is simply no big picture transparency. I'm concerned (may or may not be true but there's no way to know) that there is a relatively small numbers of block happy admins dealing in this realm, thereby chasing away potential editors and thereby harming Wikipedia (which appears to be having a major problem recruiting/retaining new editors). I'm asking someone with the tech abilities to A. create a page the logs all talkpage templated unblock requests and memorializes them over time B. have the log contain a link to each talkpage and info as to whether the request has been granted/denied/or is pending and C. create a bot to automatically update the log page....I simply do not have the skills to make this happen...Thank you for considering.. 68.48.241.158 ( talk) 20:49, 31 May 2016 (UTC)
http://www.enjoy.org.nz/ have completely redeveloped their website, breaking all the links. Could all currently links to their website please be redirected to the https://web.archive.org/web archives? Stuartyeates ( talk) 19:49, 20 April 2016 (UTC)
I started the question Category_talk:Battles_of_the_Middle_Ages#Catsort_by_bots.3F here but I think this is the best place for it. It does not only refer to Battles but many more topics. Is it possible to program/run a bot that properly catsorts the pages according to a general idea: Now: example, not a real case, Article "Battle of Prague" appears in the category "Category:Battles of Central Europe" under the B because it is not properly sorted under P. The correct sorting would be done on the article "Category:Battles of Central Europe|Prague".
In my latest contributions list you can see I did a couple of dozen pages manually but it's much too much (repetitive) work to correct it by hand. It seems not too hard to program, when you use "When Article starts with <Battle> and it is in a category of <Battles>... then recatsort the Article to <Battles>...|X" with X the word coming after "Battle of X", in the example "Prague". Similar for an article like "Treaties..." which now may appear under the "T" in any Treaties... category while it should be sorted like "Treaty of Versailles" -> "Cat:Treaties...|Versailles". I hope it's clear, I am not a programmer but as an amateur most of those things can be done by bots easily. Many articles are already properly sorted but also many are not (yet). Thanks and cheers, Tisquesusa ( talk) 23:07, 20 April 2016 (UTC)
Most of the Wiki pages do not have all of the WikiProjects tagged on their talk pages. On the basis of their categories this could be done. If a bot can tag the missing wikiprojects this saves a hell lot of time and makes it completer than it would ever become. Example: there are over about 4000 pages in the category category:Men's volleyball players and Category:Women's volleyball players that should all be tagged with {{WikiProject Volleyball}}. Is it possible to create a bot doing that? Sander.v.Ginkel ( Talk) 17:13, 5 February 2016 (UTC)
{{WikiProject Biography|sports-work-group=yes|sports-priority=}}
if it's not already there.
GoingBatty (
talk) 20:05, 5 February 2016 (UTC)
Via this way, according to the rules of User:Yobot#WikiProject_tagging, I want to make a bot request to tagg pages with WP Volleyball. I'm a member of Wikipedia:WikiProject Volleyball, and I posted the request on Wikipedia talk:WikiProject Volleyball. I also posted the request on User talk:Yobot. I created a list of all volleyball categories. I checked and delete wrong and double categories. Sander.v.Ginkel ( Talk) 18:21, 17 February 2016 (UTC)
@ Sander.v.Ginkel: I just want to make sure I understand this list. Did you list each individual category where all pages JUST in that category (not necessarily subcategories, although those may be listed separately) should be tagged? If I'm understanding you right, I'll get to work on this soon. ~ Rob Talk 23:58, 2 April 2016 (UTC)
Sander.v.Ginkel ( Talk) 19:33, 7 April 2016 (UTC)
I can't recall. Did Yobot already did this request? -- Magioladitis ( talk) 15:46, 30 April 2016 (UTC)
Hello all! I have a pretty straightforward WikiProject Banner request. I would like to ensure that {{
WikiProject Green Bay Packers}} is on all Green Bay Packers-related content. The tasks would be as follows:
1. Go through
Category:Green Bay Packers and all of its subcategories (see category tree below) and add {{
WikiProject Green Bay Packers}} to any article that is missing one.
|unref=yes
if there are no external links in the article. Parameter populates
Category:Unreferenced Green Bay Packers articles.|needs-image=yes
if there are no files in the article. Parameter populates
Category:Green Bay Packers articles needing images.|needs-infobox
if there is no infobox in the article.
Category:Green Bay Packers articles needing infoboxes.2. Go through Category:WikiProject Green Bay Packers and make sure that the following are assessed correctly:
|class=category
for all Category pages|class=template
for all Template pages|class=project
for all Wikipedia pages|class=file
for all File pages|class=redirect
for all Redirect pages (will be adding functionality in banner soon, will default to NA-class for now)|class=disambig
for all Disambiguation pages (will be adding functionality in banner soon, will default to NA-class for now)3. Go through Category:WikiProject Green Bay Packers and add the following parameters to {{ WikiProject Green Bay Packers}} (for the articles that were not edited in task 1):
|unref=yes
if there are no external links in the article. Parameter populates
Category:Unreferenced Green Bay Packers articles.|needs-image=yes
if there are no files in the article. Parameter populates
Category:Green Bay Packers articles needing images.|needs-infobox
if there is no infobox in the article.
Category:Green Bay Packers articles needing infoboxes.4. Produce a linked log (article and talk page links listed side-by-side in a numerical list) for my use to see what article talk pages are edited, make it easier to assess each article manually and to confirm the bot edits.
Here are the two categories that we would be working with:
I will work through each article talk page that is edited by the bot afterward to assess the page and ensure that the maintenance parameter tags are necessary. Please let me know if you have any questions or suggestions to improve this request. « Gonzo fan2007 (talk) @ 04:33, 2 June 2016 (UTC)
|class=project
, |class=redirect
and |class=disambig
. I appreciate the response, let me know if you have any other questions.
« Gonzo fan2007
(talk) @ 18:13, 6 June 2016 (UTC)
|unref=yes
according to the rules you laid out. Given the many ways that references can be placed in text (i.e. offline references), this is beyond the capability of an AWB bot and possibly all bots. I could compile a list of articles that do not have any ref tags in them, which you could manually go through.|needs-image=yes
according to the rules you laid out. The big challenge here is image parameters in infoboxes and the various templates that allow you to show files without typing the "File:" in the name. This is beyond the capabilities of an AWB bot.|needs-infobox=yes
to all articles that don't currently have "Infobox" somewhere in the text within a template. That shouldn't have any false positives/negatives.Categories, templates, files, etc. are auto assessed. The best option is to not define a class to allow autodetection. -- Magioladitis ( talk) 17:34, 8 June 2016 (UTC)
\.(gif|jpe?g|png|svg|tiff?)\s*[<|}\]]
, then it probably doesn't have images at all. This should be pretty trivial to check with Python, though. --
Edgars2007 (
talk/
contribs) 21:01, 8 June 2016 (UTC)
This is an archive of past discussions. Do not edit the contents of this page. If you wish to start a new discussion or revive an old one, please do so on the current talk page. |
Archive 65 | ← | Archive 68 | Archive 69 | Archive 70 | Archive 71 | Archive 72 | → | Archive 75 |
Over the years, I have come into situations where someone will slap a merge tag on an article, and then not follow through with adding a merge tag to the other article, if they even decide to mention why they want to merge the two pages at all. These tags can often remain up for years until they are removed, so I was wondering if there was a way to program a bot to remove these sorts of things, as I suspect a sizable chunk of merge request taggings are just that. This may be near impossible to do, but it would be worth looking into, if possible. Kevin Rutherford ( talk) 17:33, 21 January 2016 (UTC)
How is a bot supposed to know that there's no discussion? What if a discussion is started with a heading like aren't these the same thing or do we really need two articles. A human can figure out that those are likely merge discussions, but a bot can't. Oiyarbepsy ( talk) 14:42, 18 February 2016 (UTC)
For some reason, loads of our articles cite The Minor Planet Bulletin as Bulletin of the Minor Planets Section of the Association of Lunar and Planetary Observers, which is sometimes used as a subtitle to The Minor Planet Bulletin (see example). The title of the publication itself is, however, The Minor Planet Bulletin, and that's how the journal should be cited. So if someone could do a simple search/replace for
on the following pages
That would be much appreciated. I started doing it, but with ~230 articles to go through, this is much more suited for a bot. Headbomb { talk / contribs / physics / books} 00:20, 2 May 2016 (UTC)
|journal=The Minor Planet Bulletin, Bulletin of the Minor Planets Section of the Association of Lunar and Planetary Observers
, as on
10000 Myriostos. Is it desired to truncate these to |journal=The Minor Planet Bulletin
or leave them as-is? FWIW, I prefer the shorter version, both aesthetically and procedurally as the actual title. ~
Tom.Reding (
talk ⋅
dgaf) 23:55, 2 May 2016 (UTC)
The following should have the following replacement made
|publisher=Minor Planets Section of the Association of Lunar and Planetary Observers
). I'll post the list of changes here when I'm done, so that all these related changes are in 1 thread, for future reference. ~
Tom.Reding (
talk ⋅
dgaf) 00:55, 4 May 2016 (UTC)As of now, I have also permanently switched the journal-parameter from "Bulletin of the Minor Planets" to "The Minor Planet Bulletin" and removed the obsolete publisher-parameter. I'll retroactively amend my edits back to May 1st. Thank you all for your efforts. Rfassbind – talk 07:24, 4 May 2016 (UTC)
What would everyone think of a bot that auto-cleaned up the comments left behind by new articles created with the article wizard? For reference, I'm referring to these. -- Nathan2055 talk - contribs 04:25, 6 February 2016 (UTC)
It would be useful to have a bot that patrolled articles on fringe theories, to stop any fringe links from being added and delete passages which give undue weight. This is because there is a problem with UFO articles at the moment, many are not written according to fringe guidelines. I need someone to help me with making this, I can't code, but I want to learn how to code bots. ThePlatypusofDoom (Talk) 17:07, 12 May 2016 (UTC)
Do we have a bot that could update Wikipedia:Database reports/Long pages, say, monthly? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 21:36, 20 March 2016 (UTC)
There are hundreds and hundreds (probably thousands) of articles about sport teams (expecially football) which are completely unreferenced, and, as most of articles about sport teams contains infos like establishment/disestablishment year, honours, staff people names etc., it's very important to have references in such type of articles. So, I propose to scan WP dump for articles about sport teams, without any external link in them, without <ref>
tags, and probably without {{
Reflist}} tag and/or "References"/"Notes" section - thus list of pages to go through is formed - and then to go through them and to tag with {{
unreferenced}}. When running to do the job also it's necessary to parse source code and visual output of articles to ensure that there are no refs before tagging page. --
XXN, 18:15, 2 March 2016 (UTC)
<tt>
elementThe <tt>...</tt>
element no longer even exists in HTML. Like <font>
, it was a
separation of style and content problem. While MediaWiki doesn't choke on it, we need to stop using it like it's valid, and replace it with something more appropriate.
<span style="font-family: monospace;">...</span>
. We can't be certain that code examples were the intended use, and in many cases they were not. For the cases that are, the span markup isn't the most perfect possible markup for such examples, but using the span tag isn't "wrong", where the code tag often would be, and if anyone wants to, they can replace the span later.<code>...</code>
, as the tt dead-element is frequently used incorrectly to mark up code examples in template documentation. In the odd case that the actual output of a template uses the tt tag and is not for code (is there a template for representing telegram output? I doubt it, but it's possible), the special appearance of the code output
compared to the span output will make it obvious that the template needs to be tweaked to use a monospaced span.In special namespaces like Module and MediaWiki, uses of it should simply be identified and listed, not altered. It can safely be replaced in mainspace, Wikipedia, Portal, all talk namespaces, etc. (except where it's in nowiki or source xtags).
The bot could generate a list of templates that do not have names ending in /doc in which this substitution was performed, so they can be checked manually to ensure they don't need their ouput changed to use <span>
(or <samp>
or <kbd>
). Or just listed and not changed, and slated for manual cleanup. I'll volunteer to do that part of the cleanup, either way. —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 18:58, 19 February 2016 (UTC) Clarified. —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 21:26, 1 March 2016 (UTC)
PS: I have no objection to it being added to AWB General Fixes instead, if people think that having humans do it incrementally would be better. As long as the cleanup begins one way or another. — SMcCandlish ☺ ☏ ¢ ≽ʌⱷ҅ᴥⱷʌ≼ 19:08, 19 February 2016 (UTC)
<ref name=foo/>
(with or without quotes around the value of name=
) and <br/>
and <br>
(and the worse error "</br>
") to use proper ... />
(i.e., space, then slash, then close-angle-bracket) syntax. Preferably it would quote the value of name=
(MediaWiki doesn't choke on it, but any number of external XML parsing tools will). Another obvious fix of this sort is putting quotes around any [X|H]TML attribute values that are non-numeric, e.g. fixing class=classname
to class="classname"
(it's actually safest to quote all of them, since a numeric one could be changed to non-numeric at any time, just like any name=Johnson1999
could be changed to name=Johson 1999
by a later editor (and even MW will barf on that). The same sort of cleanup script could also perform dead-code cleanup in the form of converting any empty <element [attribute=value [...]]></element>
to <element [attribute=value [...]] />
. That last one would be arguably cosmetic and thus better done as an AWB general fix. A different kind of tool might also build up a list of pages with the same HTML id=
value used two or more times on same page, for manual fixing. I've long wondered if whatever trick is used by navboxes to detect if another navbox is present and auto-collapse could also be used to catch this error, in templates that generate id
s. Would also be nice to track down uses of <font>...</font>
(which can be auto-converted to <span>...</span>
, with some work) and <center>...</center>
(which would require manual fixing). Oh, and <acronym>...</acronym>
should in every case be changed to <abbr>...</abbr>
; the former has been invalid for years, and they support the same attributes with the same output. —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 21:26, 1 March 2016 (UTC)
On breaks: <br /> is actually only a work around for some legacy HTML parsers and is not strictly correct, though an XML parser won't choke on it (for those authors who want to serve XHTML). I doubt anything in this day and age which we support requires it. More generally, <br> is fine in Html 5, which we've been serving for years at this point. ( See StackOverflow.) So I would not approve of any change but </br> -> <br(|/| /)>. This should probably be a general fix or a CHECKWIKI fix if it isn't already.
I would support a bot to add quotation marks to any attributes, per this recent change, soon-live onwiki.
The general <element></element> -> <element/> has some issues on some legacy browsers for some attributes, related mostly to the "/" (and especially in block elements e.g. divs and paragraphs). I think these are the same legacy browsers as with the break problem.
I think the Javascript looks for the collapsible class, so you'd need to check whether "getElementByID" returns a list or a single element. Probably the latter, since an ID is supposed to be unique.
I would support replacement of font, center, and acronym, though it may be better just to template-ize them for legitimate uses and remove them for illegitimate. Probably a better task for AWB, if not AWB general fixes. -- Izno ( talk) 21:56, 1 March 2016 (UTC)
<br />
is strictly correct; XML (including XHTML) requires the /
. HTML5 does not require it, but it works fine in HTML5, and people reuse our code in more ways that we think they do, so it is best to give them the most portable code. The <br/>
(no space) format, which is also valid XML, is what certain old browsers have a problem with. It is the exact same problem as <hr/>
or any <element/>
markup, without the space, in the same browsers. The <element></element>
→ <element />
conversion does not present any problems, only <element></element>
→ <element/>
conversion. —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 13:42, 6 March 2016 (UTC)<font>
Between a bot and alot of *$(#& manual fixes, there are currently no <font>
tags in article space.<big>
But, it is not currently switched on for enwiki.</br>
There are currently no </br>
tags in articles, but alot are added everyday. Checkwiki also checks for other cases of bad <br>
tags.<tt>
to Checkwiki. I can also give a listing of articles using it via a dump file or any other tags. At one point, I was changing <tt>
to {{
mono}} or <code>
, but was getting alot of complaints, so I dropped it.
Bgwhite (
talk) 22:53, 1 March 2016 (UTC)
<big>...</big>
and <small>...</small>
would be nice to get rid of. They're convenient for entry, like <font>...</font>
, but we should not be using these things in mainspace or any equivalent, including template, book, portal, etc. It would be nice if all this stuff were purged from Wikipedia namespace, too. Really, everywhere, though I guess we care least of all if they remain in the talk spaces. People can complain all they want, but <tt>...</tt>
in particular is just dead and they have to move on with their lives. That whole "acceptance" stage of grieving. :-) —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 13:42, 6 March 2016 (UTC)
<tt>
tags from the March dump can be found at
User:Bgwhite/Sandbox1. There are sports articles at the beginning of the list in which the tt tag should be removed and not converted to <code>
. Not all tt tags should be converted to code, so a bot couldn't run on the list. Manual editing should be done.
Bgwhite (
talk) 18:51, 7 March 2016 (UTC)
{{
mono}}
in 2008 or so, because <tt>...</tt>
was frequently being used in articles. A bot or AWB script or whatever could add a silent parameter for tracking purposes. —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 18:59, 26 March 2016 (UTC)@
SMcCandlish: As usual, you've completed the work before I've even thought of it. I'd recommend placing that template around the text currently in tt tags, with additional parameter | needs_review = yes
placed at the end (or beginning, whichever's easier). Someone can create a tracking category of every transclusion with a nonempty needs_review
parameter at a later time. To be clear, this is not a substitute for properly informing editors that the edit is worth reviewing via the edit summary. A proper implementation should do both. ~
Rob
Talk 19:34, 26 March 2016 (UTC)
\<tt\>([^\<]*)\<\/tt\> --> {{mono|$1|needs_review=yes}}
on whichever namespaces this task has consensus to run on. It would miss something that had additional HTML tags within the tt tags, but it should hit most things, and the rest probably are worth editor review anyway. Where is the consensus discussion for this, by the way? ~
Rob
Talk 00:25, 31 March 2016 (UTC)Be careful about your proposed changes to the name=
tags. The danger points are, offhand 1. The possible (though horrible) situation where Johnson1999 and "Johnson1999" are different references. and 2. Accidental breaking of citation templates, e.g. the Harvard Referencing templates.
As for the rest, one issue is that this makes the code somewhat harder for editors to understand. I'd suggest using templates wherever the code gets complicated. Complicated, naked HTML is an accessibility issue. . Adam Cuerden ( talk) 00:33, 31 March 2016 (UTC)
name=foo/
is also distinct from name=foo
then I think that only the accessibility issue (templating instead of complicated raw HTML) remains. Perhaps {{
mono}}, which already exists?
Adam Cuerden (
talk) 00:43, 31 March 2016 (UTC)
<strike>...</strike>
is <s>...</s>
. WP emits a doctype of plain "HTML" (i.e., HTML5), not HTML 4.0 strict, anyway. If the argument is that any strike-through, including <s>...</s>
and <del>...</del>
is an accessibility issue, that seems to be a more theoretical matter; both those tags are part of the HTML specs, and this thread is about complying with them but getting rid of use of the invalid <tt>...</tt>
element. —
SMcCandlish ☺
☏
¢ ≽ʌⱷ҅ᴥⱷʌ≼ 09:19, 1 April 2016 (UTC)Template:Filmr and Template talk:Filmr both show up at Special:WantedPage with about 3000 incoming links each. The template itself was deleted in 2012 since it was just a holding page. It's used in many non-free screenshots as part of the fair use rationale parameter within the Template:Navbox used for the rationale. Could a bot remove the entire name parameter from these Navbox templates on the pages that are linking to Filmr? The box will then stop trying to link to the deleted template, talk and edit option as seen here and instead will be a plain box like this. Either the parameter can be removed or a bot could just as easily remove the text "Filmr" to reduce the chance of error. -- Ricky81682 ( talk) 09:26, 30 March 2016 (UTC)
{{
Navbox}}
with a custom template. All the best:
Rich
Farmbrough, 00:17, 3 April 2016 (UTC).Could somebody recreate User:Pigsonthewing/Direct calls to Infobox for me, please? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:59, 25 April 2016 (UTC)
\{\{\s*([Ii]nfo box|[Ii]nfoBox|[Ii]nfobox [Cc]onditionals|[Ii]nfobox/old|[Ii]nfobox/row|[Rr]ow)\s*[\|\}]
, which includes all of the #Redirects to {{
Infobox}} (and only #Redirects), returned these 3 additional pages:
Windows NT 4.0,
Windows Embedded Automotive,
Windows Embedded Industry. ~
Tom.Reding (
talk ⋅
dgaf) 15:38, 4 May 2016 (UTC)
WP:MOSTEDITED is sorted by both the number of editors and the number of edits, based on a formula to convert them both into standard scores from their approximately lognormal distribution as non-negative integers, and then adding the resulting normal scores together before sorting to rank the articles.
I want to do the same thing with both pageviews and ORES wp10 quality scores. The pageviews API is at e.g. https://wikimedia.org/api/rest_v1/metrics/pageviews/ ( per-article documentation) and the ORES API is at https://ores.wmflabs.org/v2/#!/scoring/get_v2_scores_context_model_revid
I want to start with, say the top 1,000 articles from May 1, 2016 and get a list of those with the lowest ORES scores, and go from there. Who can help me please? EllenCT ( talk) 14:40, 10 May 2016 (UTC)
Done as per below; shortcut WP:POPULARLOWQUALITY. EllenCT ( talk) 05:08, 23 May 2016 (UTC)
Here's what I have so far for trying to find the lowest quality high-popularity articles:
code
|
---|
import sys import urllib2 import safeJSON r1 = urllib2.Request(' https://wikimedia.org/api/rest_v1/metrics/pageviews/top/en.wikipedia.org/all-access/2016/05/15') # use about 5 days ago (test how long ago is reliable) for the date at the end of that url in production e.g.: # import datetime # print ((date.today() - timedelta(days=5)).strftime("%Y/%m/%d")) try: u1 = urllib2.urlopen(r1) top_articles = safeJSON.loads(u1.read()) for i in range(2, 999): article = top_articles['items'][0]['articles'][i]['article'] views = top_articles['items'][0]['articles'][i]['views'] title = article.replace('_', ' ') r2 = urllib2.Request(' https://en.wikipedia.org/w/api.php?action=query&format=json&prop=revisions&titles=' + article) try: u2 = urllib2.urlopen(r2) revinfo = safeJSON.loads(u2.read()) revid = revinfo['query']['pages'][revinfo['query']['pages'].items()[0][0]]['revisions'][0]['revid'] r3 = urllib2.Request(' https://ores.wmflabs.org/v2/scores/enwiki/wp10/' + str(revid)) try: u3 = urllib2.urlopen(r3) ores = safeJSON.loads(u3.read()) prediction = ores['scores']['enwiki']['wp10']['scores'][str(revid)]['prediction'] if prediction in ['Stub', 'Start', 'C']: print('[' + '[%s]] %s-class with %d views' % (title, prediction, views)) except: 0 # print(i, title, 'no ORES score') # e.g. article is too new except: 0 # print(i, title, 'can not get latest revision') # e.g. Special: namespace except: print('can not load top articles') # e.g., less than 5 or so days ago |
Can someone who understands bots please turn that into a WP:MOSTEDITED-style report updating daily with the most recent top articles date (at the end of the first url)? EllenCT ( talk) 09:30, 22 May 2016 (UTC)
@ Bamyers99: can you please add this to DataflowBot and link the daily report from WP:BACKLOG? Ideally the output would be in similar tabular format sorted first by Stub, Start, and C-class prediction, then by descending pageviews. EllenCT ( talk) 18:13, 22 May 2016 (UTC)
@ Bamyers99: would you please make the warning added to the top of the report permanent until issues with pageview transience (I think we can use a simpler version of the algorithms described here) and ORES predictions for articles composed mostly of templates get sorted out? EllenCT ( talk) 15:15, 23 May 2016 (UTC)
@ Bamyers99: please see the suggestions here for some recommended improvements to the warning message. I am trying to get a top-100,000 list so we may be able to drop the C-class predictions, which aren't generally in urgent need of improvements. EllenCT ( talk) 12:57, 25 May 2016 (UTC)
Further discussion on the bot's talk page please. EllenCT ( talk) 21:15, 27 May 2016 (UTC)
Some time ago, the parameter |licensor=
was corrected to |licensee=
, however the template calls in article space were never updated to reflect the name change. Now that |licensor=
is being phased out, the call will need to be updated.
I would have done this myself using AWB, but considered it to trivial of an edit to be allowed under AWB usage rules. — Farix ( t | c) 11:38, 21 May 2016 (UTC)
@ TheFarix: now the change is part of AWB general fixes. It is easy to be done now. -- Magioladitis ( talk) 16:07, 23 May 2016 (UTC)
Hello all! Not sure if this is the correct place to request this, so let me know if I should go elsewhere. I am working on restarting the Wikipedia:WikiProject Green Bay Packers and would like to get an updated and completed list of all of the articles that have had some type of recognition (featured articles, featured lists, DYK, etc.). I would like a query to be run to generate a list of Green Bay Packers-related content that some form of recognition. This could be done by finding searching all pages that are in Category:WikiProject Green Bay Packers and querying out those that have {{ Article history}} and {{ DYK talk}}. I am also open to other ideas that would complete the task. The main output would be a list or relevant articles (which could be placed here). Let me know if this can be done, or if I should ask elsewhere. Appreciate the help! Thanks, « Gonzo fan2007 (talk) @ 01:18, 30 May 2016 (UTC)
More than a week ago I placed my #Bot request above, but nobody responses. What's going wrong? Sander.v.Ginkel ( Talk) 09:38, 25 February 2016 (UTC)
I can start the tagging after the weekend. -- Magioladitis ( talk) 10:13, 25 February 2016 (UTC)
I've suggested before that instead of simply archiving this page, requests should be archived in one of two ways: "done/won't do/not valid" vs. "awaiting action". That way, valid requests that are not recent will be more readily found by someone who might want to work on them. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:08, 14 March 2016 (UTC)
Most well-developed Wikipedia pages of a species have information from the IUCN Red List of Threatened Species. Unfortunately there doesn't seem to be an automated process to keep the information up to date.
At the moment the data on the page for Gorilla gorilla was added in 2009. Wikipedia thus lists incorrectly "Current Population Trend" as unknown when the IUCN list it as "decreasing". I imagine that there are a lot of pages where the information that's in Wikipedia is outdated. Many smaller pages of species such as the Unicorn leatherjacket completely lack the information, while IUCN lists the information on the status of Unicorn leatherjacket (Aluterus monoceros) on their page.
Given that the IUCN reports their data in a very orderly fashion it should be straightforward to write a bot that regularly transfers new data from ICUN to Wikipedia. — Preceding unsigned comment added by ChristianKl ( talk • contribs) 14:58, 3 April 2016 (UTC)
It's been a while since we've had a tagging run the project, so if a bot could tag the following articles with {{ WikiProject Academic Journals}} this would be great. Were' talking
|class=
from other banners when possible.
|class=
should be used.|needs-infobox=yes
. Some articles won't need an infobox, but they can be dealt with one by one.|needs-infobox=yes
that have an infobox should be untagged. They can be found at
Category:Journal articles needing infoboxesAdditionally
|cover=
of {{
Infobox journal}} should be tagged with {{
WikiProject Academic Journals}}.Headbomb { talk / contribs / physics / books} 15:59, 5 April 2016 (UTC)
Headbomb I can do it. -- Magioladitis ( talk) 16:07, 5 April 2016 (UTC)
Headbomb Please leave a message at Wikipedia talk:WikiProject Academic Journals to link to this discussion here so everyone is aware about it. Thanks, Magioladitis ( talk) 16:27, 5 April 2016 (UTC)
Headbomb I know but after all these complains I keep getting I want to be on the safe side. I also still look for volunteers to do the tagging instead of me. -- Magioladitis ( talk) 20:52, 5 April 2016 (UTC)
Headbomb For instance United Nations Economic Commission for Europe contains Infobox journal. -- Magioladitis ( talk) 21:38, 5 April 2016 (UTC)
I auto-assessed for Stub/Start class and manually fixed all in NA-class. -- Magioladitis ( talk) 10:28, 7 April 2016 (UTC)
Headbomb I can't find an easy to way to fulfill the images request. -- Magioladitis ( talk) 10:33, 7 April 2016 (UTC)
I am so happy to discover that I was the one who did a previous bot tagging in 2011. 5 years ago :) -- Magioladitis ( talk) 11:21, 7 April 2016 (UTC)
The subheaders of Wikipedia:Administrators' noticeboard/Requests for closure are "Requests for comment", "Backlogs", "XfD", "Administrative", and "Requested moves". For each subheader, add in the <includeonly> section that is visible only to WP:AN (to which WP:ANRFC is transcluded) how many discussions are waiting closure in that section and how old the oldest discussion in that section is. Maybe update this once a day. Pinging BU Rob13 ( talk · contribs), who suggested this here. Cunard ( talk) 20:46, 10 April 2016 (UTC)
I write in my capacity as one of two Wikipedians in Residence at TED.
Firstly, I would appreciate some help, please, in counting links to TED talks (URLs including [Now resolved.]
/talks/
), topics (URLs including /topics/
) and speaker profiles (URLs including /speakers/
) in this and the other top ten largest Wikipedias. There is some prior discussion of the issue at
WP:VPT#External links by page type.
Secondly, it would be a good idea to clean up TED links. We have both external links and links in citations, to pages like:
and:
In such links, the index.php/
and .html
parts are redundant; and http://
would be better replaced by https://
Links may have one, two, or all three of these issues.
Thirdly, external links for TED speakers should be replaced using {{ TED speaker}} - though that might be better done once Wikidata is populated with TED IDs, and values can be called from there. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 16:57, 7 March 2016 (UTC)
Fourthly, links like https://ted.com/daphne_koller_what_we_re_learning_from_online_education should have the "www" prefix added. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 13:54, 8 March 2016 (UTC)
{{ TED talk}} is now available, also. Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 12:04, 9 March 2016 (UTC)
I've now determined that there are
148 links whose URL includes ted.com/index.php/
, and
883 links whose URL ends with .html
.
Andy Mabbett (Pigsonthewing);
Talk to Andy;
Andy's edits 11:26, 14 March 2016 (UTC)
@ Hasteur: Thank you. Yes, except:
https://
is
preferred.-- Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 14:03, 14 March 2016 (UTC)
Update: Links in the form http://www.ted.com/talks/lang/en/thandie_newton_embracing_otherness_embracing_myself.html
should also drop the lang/en/
component (for all langauge codes; 93 instances).
Andy Mabbett (Pigsonthewing);
Talk to Andy;
Andy's edits 20:36, 14 March 2016 (UTC)
As Hasteur is unavailable, can anyone else help with this, please? Andy Mabbett (Pigsonthewing); Talk to Andy; Andy's edits 15:10, 14 April 2016 (UTC)
Wikipedia Zero's made us attractive as a piracy host ( T129845). To combat music piracy I suggest setting up an audio fingerprinting system like Echoprint or AcoustID. These are open sourced (unlike Shazam or Gracenote) and supposedly easy to get running. — Dispenser 23:27, 14 April 2016 (UTC)
Can someone take over for User:Wikinews Importer Bot since Misza13 seems to have disappeared about a year ago? It would be greatly appreciated. ··· 日本穣 · 投稿 · Talk to Nihonjoe · Join WP Japan! 01:27, 17 February 2016 (UTC)
I appreciate the update. I wish you luck on finding that kind soul. ··· 日本穣 · 投稿 · Talk to Nihonjoe · Join WP Japan! 16:00, 21 April 2016 (UTC)
Hi,
Can someone please sort the names in Category:Wives of Ottoman Sultans and Category:Daughters of Ottoman sultans so that they are arranged by first name and not by title? These people don't actually have a last name (see List of Ottoman titles and appellations and a corresponding category at Turkish wikipedia) and right now their categories are chaotic, some of them are arranged by first name and some of them by title. (I expect this to be a recurring problem as some editors might not be familiar with the fact that not everything is a surname that occurs at the end of a name...) I wanted to sort them out but there are over 100 of them.
Thanks, – Alensha talk 17:12, 28 May 2016 (UTC)
Currently there exists a log of all blocks, which can be examined. The fact that it exists and can be examined is crucially important to transparency. There exists no such log of unblock requests which links to them/lists whether or not they were granted. Therefore, there is no practical manner by which to get a big picture sense of what is going on in the realm of unblock requests; there is simply no big picture transparency. I'm concerned (may or may not be true but there's no way to know) that there is a relatively small numbers of block happy admins dealing in this realm, thereby chasing away potential editors and thereby harming Wikipedia (which appears to be having a major problem recruiting/retaining new editors). I'm asking someone with the tech abilities to A. create a page the logs all talkpage templated unblock requests and memorializes them over time B. have the log contain a link to each talkpage and info as to whether the request has been granted/denied/or is pending and C. create a bot to automatically update the log page....I simply do not have the skills to make this happen...Thank you for considering.. 68.48.241.158 ( talk) 20:49, 31 May 2016 (UTC)
http://www.enjoy.org.nz/ have completely redeveloped their website, breaking all the links. Could all currently links to their website please be redirected to the https://web.archive.org/web archives? Stuartyeates ( talk) 19:49, 20 April 2016 (UTC)
I started the question Category_talk:Battles_of_the_Middle_Ages#Catsort_by_bots.3F here but I think this is the best place for it. It does not only refer to Battles but many more topics. Is it possible to program/run a bot that properly catsorts the pages according to a general idea: Now: example, not a real case, Article "Battle of Prague" appears in the category "Category:Battles of Central Europe" under the B because it is not properly sorted under P. The correct sorting would be done on the article "Category:Battles of Central Europe|Prague".
In my latest contributions list you can see I did a couple of dozen pages manually but it's much too much (repetitive) work to correct it by hand. It seems not too hard to program, when you use "When Article starts with <Battle> and it is in a category of <Battles>... then recatsort the Article to <Battles>...|X" with X the word coming after "Battle of X", in the example "Prague". Similar for an article like "Treaties..." which now may appear under the "T" in any Treaties... category while it should be sorted like "Treaty of Versailles" -> "Cat:Treaties...|Versailles". I hope it's clear, I am not a programmer but as an amateur most of those things can be done by bots easily. Many articles are already properly sorted but also many are not (yet). Thanks and cheers, Tisquesusa ( talk) 23:07, 20 April 2016 (UTC)
Most of the Wiki pages do not have all of the WikiProjects tagged on their talk pages. On the basis of their categories this could be done. If a bot can tag the missing wikiprojects this saves a hell lot of time and makes it completer than it would ever become. Example: there are over about 4000 pages in the category category:Men's volleyball players and Category:Women's volleyball players that should all be tagged with {{WikiProject Volleyball}}. Is it possible to create a bot doing that? Sander.v.Ginkel ( Talk) 17:13, 5 February 2016 (UTC)
{{WikiProject Biography|sports-work-group=yes|sports-priority=}}
if it's not already there.
GoingBatty (
talk) 20:05, 5 February 2016 (UTC)
Via this way, according to the rules of User:Yobot#WikiProject_tagging, I want to make a bot request to tagg pages with WP Volleyball. I'm a member of Wikipedia:WikiProject Volleyball, and I posted the request on Wikipedia talk:WikiProject Volleyball. I also posted the request on User talk:Yobot. I created a list of all volleyball categories. I checked and delete wrong and double categories. Sander.v.Ginkel ( Talk) 18:21, 17 February 2016 (UTC)
@ Sander.v.Ginkel: I just want to make sure I understand this list. Did you list each individual category where all pages JUST in that category (not necessarily subcategories, although those may be listed separately) should be tagged? If I'm understanding you right, I'll get to work on this soon. ~ Rob Talk 23:58, 2 April 2016 (UTC)
Sander.v.Ginkel ( Talk) 19:33, 7 April 2016 (UTC)
I can't recall. Did Yobot already did this request? -- Magioladitis ( talk) 15:46, 30 April 2016 (UTC)
Hello all! I have a pretty straightforward WikiProject Banner request. I would like to ensure that {{
WikiProject Green Bay Packers}} is on all Green Bay Packers-related content. The tasks would be as follows:
1. Go through
Category:Green Bay Packers and all of its subcategories (see category tree below) and add {{
WikiProject Green Bay Packers}} to any article that is missing one.
|unref=yes
if there are no external links in the article. Parameter populates
Category:Unreferenced Green Bay Packers articles.|needs-image=yes
if there are no files in the article. Parameter populates
Category:Green Bay Packers articles needing images.|needs-infobox
if there is no infobox in the article.
Category:Green Bay Packers articles needing infoboxes.2. Go through Category:WikiProject Green Bay Packers and make sure that the following are assessed correctly:
|class=category
for all Category pages|class=template
for all Template pages|class=project
for all Wikipedia pages|class=file
for all File pages|class=redirect
for all Redirect pages (will be adding functionality in banner soon, will default to NA-class for now)|class=disambig
for all Disambiguation pages (will be adding functionality in banner soon, will default to NA-class for now)3. Go through Category:WikiProject Green Bay Packers and add the following parameters to {{ WikiProject Green Bay Packers}} (for the articles that were not edited in task 1):
|unref=yes
if there are no external links in the article. Parameter populates
Category:Unreferenced Green Bay Packers articles.|needs-image=yes
if there are no files in the article. Parameter populates
Category:Green Bay Packers articles needing images.|needs-infobox
if there is no infobox in the article.
Category:Green Bay Packers articles needing infoboxes.4. Produce a linked log (article and talk page links listed side-by-side in a numerical list) for my use to see what article talk pages are edited, make it easier to assess each article manually and to confirm the bot edits.
Here are the two categories that we would be working with:
I will work through each article talk page that is edited by the bot afterward to assess the page and ensure that the maintenance parameter tags are necessary. Please let me know if you have any questions or suggestions to improve this request. « Gonzo fan2007 (talk) @ 04:33, 2 June 2016 (UTC)
|class=project
, |class=redirect
and |class=disambig
. I appreciate the response, let me know if you have any other questions.
« Gonzo fan2007
(talk) @ 18:13, 6 June 2016 (UTC)
|unref=yes
according to the rules you laid out. Given the many ways that references can be placed in text (i.e. offline references), this is beyond the capability of an AWB bot and possibly all bots. I could compile a list of articles that do not have any ref tags in them, which you could manually go through.|needs-image=yes
according to the rules you laid out. The big challenge here is image parameters in infoboxes and the various templates that allow you to show files without typing the "File:" in the name. This is beyond the capabilities of an AWB bot.|needs-infobox=yes
to all articles that don't currently have "Infobox" somewhere in the text within a template. That shouldn't have any false positives/negatives.Categories, templates, files, etc. are auto assessed. The best option is to not define a class to allow autodetection. -- Magioladitis ( talk) 17:34, 8 June 2016 (UTC)
\.(gif|jpe?g|png|svg|tiff?)\s*[<|}\]]
, then it probably doesn't have images at all. This should be pretty trivial to check with Python, though. --
Edgars2007 (
talk/
contribs) 21:01, 8 June 2016 (UTC)