This is an archive of past discussions. Do not edit the contents of this page. If you wish to start a new discussion or revive an old one, please do so on the current talk page. |
Archive 1 | Archive 2 | Archive 3 | Archive 4 |
Hi,
Per earlier discussion at Wikipedia talk:Naming conventions (geographic names)#Serbia village name format, we have a discrepancy in the formatting of disambiguation markers for villages in Serbia. Can someone please help produce a list of all articles in the main namespace that are underneath any of the categories of Category:Populated places in Serbia and match the regular expression /\(.*\)$/? We'd then be able to redact the list to make sure we don't have any false positives or negatives, and put this up for a mass move.
Likewise, if the query isn't too taxing on the available resources, we could do the analogous thing with articles underneath Category:Populated places in Bosnia and Herzegovina that have the same issue (it would be great to see what the impact of the mass move could be there).
TIA! -- Joy [shallot] ( talk) 07:21, 1 April 2021 (UTC)
.*\(.*\)$
. I put the category depth to 10, you can experiment with going deeper to make sure that is deep enough to include everything. I see a lot of non-villages mixed into the data... I think some of the deep sub-categories and/or their articles may need attention. For example,
Probus (emperor) is in the results because he is included in
Category:Sirmium, which was a Roman city that happened to be located in Serbia. Perhaps an additional category or a different RegEx would help to prune out these non-villages. –
Novem Linguae (
talk) 07:53, 1 April 2021 (UTC)
Chlod and I worked on this query, but both our attempts timed out.
Here's the code. Any suggestions for speeding this up? I think the revisions table has 1 billion rows, so it's quite a large table. Also, I read somewhere that WHERE, GROUP BY, and ORDER BY should be run on columns that are indexed/keyed for faster queries, but currently all 3 of those are being run on non-keyed columns. The
SQL optimizer tool is giving Query plan 1.1 is using filesort or a temporary table. This is usually an indication of an inefficient query. If you find your query is slow, try taking advantage of available indexes to avoid filesort.
–
Novem Linguae (
talk) 17:09, 2 April 2021 (UTC)
My approach was to check for pages in Category:All articles needing additional references. I tried Petscan, but couldn't find a filter for created date. I tried Quarry, and got it mostly working, but I got stuck on filtering by created date. I'm surprised there isn't a "created date" field in the `pages` table. Would appreciate any help. Even if it's to smack me for missing a simple solution :-) – Novem Linguae ( talk) 12:16, 12 April 2021 (UTC)
10+ years ago, as a first cutoff? For Wikipedia:Village pump (proposals)#Make page movers eligible to move move-protected pages – xeno talk 21:33, 26 April 2021 (UTC)
WHERE MAX(log_timestamp) < DATE_FORMAT(DATE_ADD(NOW(), INTERVAL -10 YEAR), "%Y%m%d%H%i%s")
, but it did not like WHERE MAX
, per
[1]. Again, if anyone sees a good solution, feel free to fork and post. Thanks. –
Novem Linguae (
talk) 22:41, 26 April 2021 (UTC)
Thomas Ranch is about to be deleted, having existed in Wikipedia for over fifteen years without a single inline reference, and without ever having an external link to an independent source. Grub Smith, though a bit longer, has been in a similar state for an equally long time. I happened to come across these two not while searching for suspect articles but while doing general cleanup around the given name Thomas and the surname Smith. How many more articles are out there in this condition? I propose that the best way to find out, if technically feasible, would be to generate a list of articles that have never had an inline ref tag over the course of their existence, sorted by age, and push through them from the oldest on forward. If someone has the Wiki-fu to generate such a list, please have at it. BD2412 T 02:42, 5 April 2021 (UTC)
I am thinking that the most immediate cases can be caught by an inquiry further narrowed to articles categorized in the Category:Companies category tree, or articles categorized in Category:Living people. The living people category will likely work well, but I found that digging down through the company category tree produces mostly dirt after just a few levels. I think you'll have more success grabbing articles that have a Company or company infobox (there are two of them, case sensitive). This set of articles is very clean. Sam at Megaputer ( talk) 23:50, 8 April 2021 (UTC)
In 2019, I
wrote a bot that added {{
unreferenced}}
to 10,000 articles. It could have added many more, but the community was dubious/conservative during an RfC and I had dial it back to the most extreme cases only. Thus any external link anywhere on the page disqualified it from being tagged, including a single external link in the external link section or in the infobox transcluded from a template. After it ran, there were zero complaints of bad tagging. It is capable of distinguishing sections and main body. It can create lists instead of tagging. It scours the entire 6+ million corpus looking for candidates. This is a non-trival bot even though it sounds easy-ish there are endless edge case exceptions - for example what counts as a dab vs. article? Not at all clear when you look at it. And so many external link templates to look for. It can't be done with a query it requires a purpose built application. The suggestion of looking for the worse of the worse is interesting, but I suspect the majority of articles tagged with {{
unreferenced}}
have always been that way. --
Green
C 02:33, 9 April 2021 (UTC)
In a discussion at the WP:Core Contest, I suggested a list of the most viewed Wikipedia articles that are start class or less would be beneficial for the competition, as presumably articles of interest to swathes of readers would be important ones to get right for our reputation. Obviously it may be useful to weed out the articles that are only prominent due to recent news stories (with a sharp spike in views e.g. Ron DeSantis ( views), in favour of those that have had more consistent page views through it's life on Wikipedia. I guess we could choose a certain standard deviation for page views to fall within from day to day? Note: Sometimes an article as different Class ratings from different Wikiprojects. Can you code in that the highest rating by the various Wikiprojects is the correct one? (Sometimes it is re-classed up but the other Wikiprojects don't catch up) I wouldn't worry about the Article Important class too much as many of these may not have been classed correctly.
-- Coin945 ( talk) 15:42, 29 May 2021 (UTC)
Hi, why isn't this working? I'm trying to generate a list of all indefinitely creation-protected articles at a sysop level. Anarchyte ( talk) 10:09, 4 June 2021 (UTC)
Thanks, Cryptic — that works perfectly. Anarchyte ( talk) 11:38, 4 June 2021 (UTC)
You are invited to join the discussion at Wikipedia talk:WikiProject Sweep § June 2021 follow-up. {{u| Sdkb}} talk 21:48, 4 June 2021 (UTC)
sub1
, such as filtering out disambig pages that are making up much of the output. –
SD0001 (
talk) 07:47, 20 June 2021 (UTC)
Hello!
The SQL code in c:Category talk:Animated PNG didn't work. Does anyone know how to fix it? The goal is to find all animated PNGs. Jonteemil ( talk) 18:17, 3 September 2021 (UTC)
Hello, this is a request I made on the talk page of the Database reports page, but it hasn't been answered over there. It is part of the overall proposal discussion I have started on the WikiProject Templates to create an Unused Templates Task Force to deal with the backlog of unused templates.
I've requested four reports per the original discussion and I'm going to relist them here:
I would like the report to run for at least two months as the task force is currently in an idea stage. When the reports are going to expire, if possible I would like to be notified of when it will happen. I need four reports from the Unused Templates database:
1) All unused templates that are not stubs or redirects.
2) All stub templates listed as unused. According to one of the users on the talk page discussion, there are either exactly or about 1,000 stub templates.
3) All redirects listed as unused. According to one of the users on the talk page discussion, there are either exactly or about 69,000 redirects.
4) Templates that were created and/or edited in the last year and the present year. -- WikiCleanerMan ( talk) 13:23, 28 September 2021 (UTC)
I'd like to refine this list (articles subcategorized under Category:United States to three levels that do not already have {{ Use mdy dates}} or {{ Use dmy dates}}) to remove all articles that are subcategorized under a different country. It'd also be nice to remove all articles that contain a date in DMY format outside of a reference. Would one or both of those be possible? {{u| Sdkb}} talk 19:08, 18 October 2021 (UTC)
Articles related to Canada may use either format with (as always) consistency within each article. We shouldn't be enforcing a standard date format on Canadian articles against the MOS, and against the fact that Canada doesn't exclusively use mdy dates. Joseph 2302 ( talk) 15:03, 19 October 2021 (UTC)
Your count for Linguistic Association of Canada and the United States is right (if "for" was a typo for "four"), but your route is off - Category:Education in Canada isn't a direct subcat of Category:Canada, and Category:Academic organizations based in Canada isn't a direct subcat of Category:Educational organizations based in Canada. It goes Category:Wikipedia categories named after countries (depth 0) > Category:Canada (depth 1) > Category:Organizations based in Canada (depth 2) > Category:Organizations based in Canada by subject (depth 3) > Category:Academic organizations based in Canada (depth 4).
What depth for Category:Military of the United States? (Just articles in that category is depth 0, articles in its direct subcat Category:American military aviation is depth 1, etc. to be sure we're talking about the same thing.)
The query excluding articles in depth 4 subcats of
Category:Wikipedia categories named after countries looks like it's going to die. Not all that surprising; depth 4 from
Category:United States is 8165 categories, depth 3 from
Category:Wikipedia categories named after countries is 39249, and depth 4 from it is 166700. —
Cryptic 21:24, 19 October 2021 (UTC) Of course it finished immediately after I gave up and saved the edit - 172888 results at depth 4 without doing anything with the Military subtree;
Linguistic Association of Canada and the United States and
Party for the Commonwealth of Canada correctly excluded, the French ship articles are still there. —
Cryptic 21:29, 19 October 2021 (UTC)
Hi. I'm looking for two related queries, one yielding a count, the other a handful of sample pages:
Extra credit: which red Draftspace pagename has the highest number of mainspace in-links, and how many? Thanks! Mathglot ( talk) 06:02, 25 October 2021 (UTC)
Basically, I'm looking for things like
{{cite journal ... |journal=Foobar |publisher=Foobar}}
{{cite magazine... |magazine=Foobar |publisher=Foobar}}
{{cite journal ... |work=Foobar |publisher=Foobar}}
{{cite web ... |website=Foobar |publisher=Foobar}}
{{cite news... |newspaper=Foobar |publisher=Foobar}}
{{citation ... |journal=Foobar |publisher=Foobar}}
E.g. if in a template you have any of [|journal=
, |magazine=
, |work=
, |website=
, |newspaper=
] that matches a |publisher=
in the same template, I'd want to know about it. It's OK if whitespace and casing differs. Thanks.
Headbomb {
t ·
c ·
p ·
b} 00:51, 17 October 2021 (UTC)
So last night my time I found this IP range with a rather interesting edit summary pattern, and thought I might try to use my limited MySQL knowledge to figure out how to use database queries to search for more instances of it. I made a working query for the RecentChanges table using query 24762 but my attempt to apply it to the whole revision table failed spectacularly (it seems to have hung (it's been running for about twelve hours now) but I can't figure out how to stop it). I originally tried searching for the phrase "%/info/en/?search=American_and_British_English_spelling_differences#/media/File:Defence_Defense_Labour_Labor_British_American_spelling_by_country.svg%" but this thread on mediawiki.org said that prefix searches like this are unbelievably slow, so I changed the query string to "China uses US spelling%" and simplified the query as far as I could without success. I can't figure out how to make use of the "revision_comment" alternative view in this case. Judging by the RecentChanges test and other independent searching of IP ranges that I've done, I'm not sure there'll be many more results, but if anyone can help me make a successful query for this info in the revision table, I'd appreciate it. Graham 87 08:31, 7 November 2021 (UTC)
Related to this idea, I'd like to create a bunch of redirects from e.g. Jane Lastname to Jane Q. Lastname. Would it be possible to compile a list of articles that meet the following criteria?
[word1] [capitalletter]. [word2]
[word1] [word2]
, nor any other articles with the title [word1] [anything] [word2]
Cheers, {{u| Sdkb}} talk 22:24, 21 November 2021 (UTC)
I
drafted a query for this but ran into some issues. Can you guys take a look and see if you can fix it? I'm trying to use the alias day
in some spots that it doesn't like, and I don't know how to fix it. Thanks. –
Novem Linguae (
talk) 02:18, 13 December 2021 (UTC)
I wrote a
query that gets an article's author. If I wanted to get this to get the authors of multiple articles in one query, what's the best way to refactor it? WHILE? Subquery? Here's my list of page IDs. Thank you. 66411662, 63051276, 67331300, 67494602, 67251604, 67475738, 67025524, 67282385, 67492243, 67505824, 68570713, 65754113, 68673796, 67481281, 68288689, 67485694, 68624634, 67062564, 67486327, 65912571, 67495969, 65558215, 67162967, 67504737, 66978220, 65952291, 67306801, 64208962, 67222236, 67365517, 68510913, 67480539, 66411751, 65228882, 67252944, 66476730, 68469744, 67008083, 66555751, 67282708, 67419043, 68693806
–
Novem Linguae (
talk) 16:57, 27 December 2021 (UTC)
I have a couple of stuck queries ( 61115, 58635) which still claim to be running hours beyond the max_statement_time I set. I'm aware that the tables are large, but I'm an experienced Quarry user who used to write SQL for a living, and the optimizer suggests that the queries use the expected indexes efficiently. The Stop button returns an HTML 500 response and seems to do nothing. Are these SQLs effectively dead, or is there some sort of regular clean-up process to release them? I'm reluctant to fork them or otherwise repeat my work in case it overloads the server. I remember that queries used to claim to be running after they had died for some reason (out of memory?) but in those cases it was possible to amend and re-run the query rather than it remaining permanently locked as "running". Certes ( talk) 00:29, 31 December 2021 (UTC)
Hello!
I would like a query of all files with MIME type image/x-bmp
and image-x-xcf
, respectively, across all wikis (all language versions of Wikipedia, Commons, Wikibooks, Mediawiki etc.). They can all be found at at
Special:MIMESearch/image/x-bmp and
Special:MIMESearch/image/x-xcf respectively on each wiki (or
Special:Search/file: local: filemime:bmp /
Special:Search/file: local: filemime:xcf). I would like the list to be rendered like this so a wikilink will work on any wiki:
w:File.NAME.bmp
fr:File.NAME.bmp
c:File:NAME.bmp
pt:v:File:NAME.bmp
etc.
Is that possible? To clarify I would like two queries, one per MIME type. Thanks! Jonteemil ( talk) 21:28, 9 January 2022 (UTC)
I need to check these article. Can anyone do this?-- 28 ( talk) 07:52, 10 January 2022 (UTC)
Can I have a list of all mainspace and Wikipedia namespace pages containing either spelling of the word "colo(u)r", or derivative words, regardless of redirect/disambiguation status, where there is no page with the other spelling? Animal lover 666 ( talk) 09:57, 13 February 2022 (UTC)
I'm trying to select (mainspace, non-redirect, non-dab) pages which have small gaps between their page_random value and the next-smallest page_random value (again, among mainspace non-redirect non-dab pages). I have a query here, which works when limited to a small subset of pages, but scales very badly as I relax the range of page_random values. I'd very much appreciate it if anyone could suggest any optimizations that would let me run it over the whole pages table (or at least larger chunks of it). Colin M ( talk) 18:35, 4 March 2022 (UTC)
This is purely for curiosity (or a possible Signpost feature), so take it up only if it's of interest. But I'd like to create a map of the world that shows how many good articles each country has per capita, and a similar one for FAs. The query for this would be articles in Category:Good articles (or Category:Featured articles) whose Wikidata item has country (P17) set to a given country. Those would then be adapted to shade an interactive map similar to Wugapodes' {{ Interactive COVID-19 maps}}. {{u| Sdkb}} talk 08:06, 17 March 2022 (UTC)
And here's for GAs with people, a larger data set:
Graphs are unavailable due to technical issues. There is more info on Phabricator and on MediaWiki.org. |
The winner is still Greenland, but Iceland's exceptional performance is looking less like a fluke. I wonder why that is (and Norway, too). And I expected India to do badly, but... geez. The only two visible countries to appear to have no GAs are Turkmenistan and Guinea-Bissau; I'll go offer rewards for those. {{u| Sdkb}} talk 19:17, 17 March 2022 (UTC)
Courtesy link (pushpin map): VPT Archive 196. Mathglot ( talk) 04:02, 2 April 2022 (UTC)
Hello, I'm interested in creating a list of orphans and articles with very few inlinks, which may be helpful as a To-do list for WikiProjects. For starters, and as an illustration and real-world example, I'm interested in a list of orphan articles that are categorized in WP:WikiProject Military history ("orphan" defined as an article with no in-links from other pages in mainspace; "in WikiProject X" means the associated talk page contains template {{ WikiProject X}}, or one of its redirects.) I was able to find this advanced search which is not quite the same thing but the "1,259" result may be a lower bound for the number of actual orphans.
Before generating any possibly long lists of articles, I'd like to first get a feeling for the scope of the problem. So, as a first db query can we do a summary query which returns eleven rows containing two values, where column one (let's call that 'nbr of inlinks') contains the integers 0 to 10, and column two ('article count') is the number of articles (ns=0) in WP:WikiProject Military history containing exactly that number of inlinks from other mainspace articles. So, e.g., the top data row contains a '0' on the left, and the number of orphans in WP:MILHIST on the right. The next row, a '1', and the count of MILHIST articles with exactly one (non-self) in-link from mainspace. And so on.
If the article count is not too huge, could we do a second query with one column containin the list of orphan articles in MILHIST as well? Please skip redirects, and I'm probably forgetting some other stuff, but anything that wouldn't be useful in a "ToDo" artile list for a wikiproject should be skipped as well. If you can request the article titles as a wikilink, that would be handy, and if it's not too difficult to add a second column containing a wikilinked Q-item from WikiData for those articles that have Wikidata links with at least one non-English Wikipedia link (i.e., >= 1 link excluding en-wiki and simple), that would be ideal. (Example: Olga Bogdanova is an orphan, but has d:Q26243976 attached to it, containing two non-English Wikipedia articles (ar, and ru); so in this case, col. 2 would be d:Q26243976. Thanks! Mathglot ( talk) 02:02, 2 April 2022 (UTC)
WITH RECURSIVE
in my
I-Spy book of SQL keywords. Analysis by link count is 0: 1540, 1: 5586, 2: 8008, 3: 9115, 4: 8892, 5: 7756, 6: 6817, 7: 6045, 8: 5156, 9: 4545, 10: 4104.
Certes (
talk) 23:32, 2 April 2022 (UTC)
This was about more than mere curiosity, and I hope it will lead to significant improvements in orphan backlog reduction. I've created a pilot at WikiProject Military history, and if you're interested in following along, see Wikipedia talk:WikiProject Military history#De-orphaning articles, and follow the link to the demo page. Mathglot ( talk) 01:01, 4 April 2022 (UTC)
As bots can't fix these. Thanks! 🐶 EpicPupper (he/him | talk) 21:50, 15 April 2022 (UTC)
AND pr_level IN ('sysop', 'interface-admin')
to the end of the first query, before the semicolon. I didn't mainly to prove that there aren't any interface-admin-protected ones, as opposed to a typo in the group name, or a missed rename like the recent 'oversight' to 'suppress'. (I'm fairly sure the current configuration disallows explicit protection to interface-admin anyway, and if we haven't changed the internal name from sysop to administrator yet we aren't likely ever to, but might as well be sure.) —
Cryptic 22:48, 15 April 2022 (UTC)Because of discussion at Help talk:Citation Style 1 § |location= without |publisher=, I wonder if there a way to get the number of articles that meet these criteria:
{{
cite book}}
template that has:
|location=<location>
AND
|publisher=
is empty OR|publisher=
is omitted{{
citation}}
template that:
|journal=<journal title>
OR|magazine=<magazine title>
OR|newspaper=<newspaper title>
OR|periodical=<periodical title>
OR|website=<website title>
OR|work=<work title>
|location=<location>
AND
|publisher=
is empty OR|publisher=
is omittedSure, we can modify Module:Citation/CS1 to add templates that match these criteria to a category but it might be better to know beforehand the magnitude of the problem if there is one.
All I want is the numbers. Thanks.
— Trappist the monk ( talk) 15:37, 23 April 2022 (UTC)
\s
does not match whitespace characters. For example,
this search does not find thousands of articles using {{
cite book}}
(with a space) but
this search does. Similarly, \w
is just an escaped w
... Rewriting your regex like this:
hastemplate:"cite book" insource:/\{ *cite book *(\| *(([^p]|p[^u]|pu[^b])[^\|\}]*|publisher *= *))*\| *location *= *[A-Za-z][^\|]*(\| *(([^p]|p[^u]|pu[^b])[^\|\}]*|publisher *= *))*\}\}/i
{{
cite book}}
with |location=
and without |publisher=
but, alas, all {{cite book}}
templates in
U.S. state, though returned by the search, have both |location=
and |publisher=
.Would it be possible to take the list here and filter it to only contain entries that have the wikitext "==Plot==" or "== Plot =="? {{u| Sdkb}} talk 02:23, 28 April 2022 (UTC)
Hi, I'd like to know if I can find out how many times a particular edit summary from a popular semi-automated tool has been used. Ohconfucious's MOS:NUMDATES script's former edit summary was: date formats per [[MOS:DATEFORMAT]] by [[WP:MOSNUMscript|script]]
(date formats per
MOS:DATEFORMAT by
script).
Dawnseeker2000 21:30, 8 May 2022 (UTC)
Hello, folks,
Is it possible to run a query on my own contributions and logs? This is not for a higher purpose, I am just curious about how many Teahouse invitations I've posted (I think it's in the tens of thousands) so I'd run a query to see how many of my edit summaries state "Welcome to Wikipedia: check out the Teahouse!" since that is the default edit summary when posting an invitation. I'm also curious about the pages that I've deleted, if I could run a query on the deletion summaries on Special:Log/Liz to see the proportion that were based on a particular speedy deletion criteria or whether they were deleted PRODs or because of an AfD decision. This query might be more complicated because sometimes multiple CSD criteria are listed in a deletion summary or the wording is not concise. But I don't think I've ever left a blank deletion summary so there would be some reason listed for every page deletion I've made.
If these tools are really intended to help manage the project, I understand that my request is purely personal. It's just that my edit count is growing and I know at a certain point in the future, Edit Count will no longer be available for me to see my how my edits breakdown so I thought I'd inquire about these two queries before my edit counts and log entries become too large for these to be practical requests.
Many thanks for any feedback you have, even if you tell me that these results are not possible to obtain. I'm constantly in awe of editors and admins who create bots or write scripts and make tools which make our editing here so much easier and simple. Your talents are appreciated! Thanks again. Liz Read! Talk! 20:24, 9 May 2022 (UTC)
Hey, I'm trying to list French sub talkpages 'À_faire' (like fr:Discussion:Dantès/À faire) that still exists while talkpage (like fr:Discussion:Dantès) OR main page (like fr:Dantès) doesn't exist.
I started with :
SELECT page_title
FROM page talkpage
WHERE talkpage.page_title LIKE '%/À_faire'
AND talkpage.page_namespace IN (1,5,11,15,101,119,711,829)
But, I'm not able to end this SQL query with NOT EXISTS
in order to check :
Maybe this is a starting point :
AND NOT EXISTS (
SELECT 1
FROM page mainpage
WHERE mainpage.page_namespace=talkpage.page_namespace-1
AND mainpage.page_title=talkpage.page_title)
But... it's going to list French sub talkpages 'À_faire' WHEN pages like fr:Dantès/À faire doesn't exist BUT that's not what I'm looking for. I need to check IF fr:Dantès OR fr:Talk:Dantès doesn't exist instead. Any clue?
Thank's -- LD ( talk) 22:10, 12 May 2022 (UTC)
AND NOT EXISTS (
SELECT 1
FROM page mainpage
WHERE mainpage.page_namespace IN (talkpage.page_namespace-1, talkpage.page_namespace)
AND mainpage.page_title=SUBSTRING(page_title,1,CHAR_LENGTH(page_title)-8))
CHAR_LENGTH('/À_faire')
. Did you also want to see pages where one of the two exists but not the other?
Certes (
talk) 22:50, 12 May 2022 (UTC)
AND (NOT EXISTS (SELECT 1 FROM page mainpage
WHERE mainpage.page_namespace = talkpage.page_namespace-1
AND mainpage.page_title=SUBSTRING(page_title,1,CHAR_LENGTH(page_title)-8))
OR NOT EXISTS (SELECT 1 FROM page mainpage
WHERE mainpage.page_namespace = talkpage.page_namespace
AND mainpage.page_title=SUBSTRING(page_title,1,CHAR_LENGTH(page_title)-8)))
Hello!
I'd like to find files whose extension don't match the MIME. I can find them using CirrusSearch, for example Special:Search/file: local: filemime:png -intitle:/\.png/i. You must however do this MIME by MIME - you can not do it all MIMEs simultaneously. It'd be nice to do a collective query of ALL files which don't match their MIME. You can find all possible file extensions and MIMEs on Special:MediaStatistics. Jonteemil ( talk) 19:58, 17 May 2022 (UTC)
Hi, I'm trying to re-run a bunch of queries on Quarry that I created a while back for looking for candidates to invite to New Page Patrol. For example: https://quarry.wmcloud.org/query/42657 among others. However, when I try to re-run them, it says "Bad database name". Could you indicate what needs to be changed to get them working again? — Insertcleverphrasehere( or here)( or here)( or here) 21:26, 26 May 2022 (UTC)
enwiki_p
in the "Enter the DB name here..." field between the description and the actual query
* Pppery *
it has begun... 21:29, 26 May 2022 (UTC)
Part 1: Need a query that gets a list of PageTriage/NPP unreviewed articles. This is tricky because I don't think there's a field anywhere that stores if an article is unreviewed. Rather, you have to JOIN logging
or JOIN pagetriage_log_table
and look for the absence of a record. Might also have to limit it by date, since PageTriage logs appear to be self-deleting after 12 months, so the absence of a log after 12 months = reviewed.
More info.
Part 2: Need a way to figure out if the unreviewed articles are species article or not. One of the best ways is to search the article text for "is a species of", but we can't search article text in SQL. Another way might be to check for any
biological sciences stub template. Another way might be to check for any article title that is in the format /[A-Z][a-z]+ [a-z]+/
(two words, first character is a capital letter, all other characters are lowercase) and I can do some manual filtering after that. –
Novem Linguae (
talk) 21:44, 26 May 2022 (UTC)
SELECT COUNT(*) FROM pagetriage_page JOIN page ON page_id=ptrp_page_id WHERE ptrp_reviewed=0 AND page_namespace=0 AND page_is_redirect=0;
at 14159. —
Cryptic 00:17, 27 May 2022 (UTC)
As suggested by
Sdkb
here, looking for a query that lists all articles in
Category:All Wikipedia level-5 vital articles but not in
Category:Articles with short description. Thank you! (I'm not
watching this page – please use {{
reply to|PerfectSoundWhatever}}
on reply) —
PerfectSoundWhatever (
t;
c) 15:01, 1 June 2022 (UTC)
Is it possible to obtain a list of recently updated talk pages efficiently, e.g. through SQL via page.page_latest? It could be a list of pages updated since a certain date, or a complete list of all pages with the date of last update, or possibly a list of all pages sorted by last update ranking. Per discussion Wikipedia:Bot_requests#Add_Reflist-talk_to_talk_pages_with_references. Green C 16:21, 3 June 2022 (UTC)
page_namespace in (1, 11)
shows just (article) Talk: and Template_talk:. I've excluded redirects: including them would speed the query up but produce lots of false positives.
Certes (
talk) 23:52, 3 June 2022 (UTC)
An edit request was made at
m:Module talk:Template translation where
Verdy p said that the language specified in the metadata of a page may not be the same with the last part of its title. For example,
m:User:Tangotango/Whatredirectshere/sv is expected to be in Swedish, but MediaWiki doesn't think so. I have tried writing a query myself using
this manual but page_lang
always return NULL
.
NguoiDungKhongDinhDanh 03:42, 15 June 2022 (UTC)
page_lang
means that page's language is similar to that of the project, then I'm looking for pages with at least one slash, without page_lang
set, where the part after the last slash isn't en
but matches the page_lang
of at least one other page. Honestly, I don't expect 1000, let alone 1.7 million.
NguoiDungKhongDinhDanh 07:42, 15 June 2022 (UTC)
page_lang
seems to be inconsistent with {{
PAGELANGUAGE}}. For example,
m:Main Page/es is clearly a Spanish translation (
expand). This request is about to be out of scope for this page so perhaps I'll need to ask at
WP:VPT. Thanks a lot anyway.
NguoiDungKhongDinhDanh 13:20, 15 June 2022 (UTC)I need these pages, but I don't know how to write a SQL query. Can anyone help me?-- Q28 ( talk) 14:21, 18 June 2022 (UTC)
I don't know if this is a query exactly, but I thought I would start here. What I am hoping to have generated is a table, a histogram of sorts, showing the number of articles that have short descriptions, put into bins by the length of the short description. Something like this:
Length (in characters) | Number of articles |
---|---|
N/A (no SD) | 2,000,000 |
1 | 10 |
2 | 100 |
3 | 265 |
... | ... |
99 | 2,204 |
100+ | 245 |
The purpose of this data request is to inform an ongoing conversation about the appropriate length of short descriptions.
Redirect pages should be excluded from the counts. The maximum bucket value should be "100+", since there should be only a hundred or so longer than that. One tricky bit is that some short descriptions are assigned by infobox templates, but I think that should be OK, since the descriptions are still listed under "Local description" on the Page information page for the article. According to "state of the project" data, there should be about 4.5 million articles with short descriptions and 1.6 million without them. The histogram does not need to be up-to-the-minute, so if it based on a database dump, that is fine. Please let me know if you have any questions. Thanks. – Jonesey95 ( talk) 03:19, 24 June 2022 (UTC)
For each AfD closed as redirect, how many times has the page been restored to an article and reverted to a redirect since then?
Example: WP:Articles for deletion/Children of This Earth, Special:PageHistory/Children of This Earth: 1 restore, 1 revert
Notes:
I think I may be able to figure this out if pointed in the right direction. Thanks! Flatscan ( talk) 04:21, 27 June 2022 (UTC)
I am looking for pages outside of Template namespace that have multiple transclusions, maybe 5 or more, and also have Linter errors. Linter errors, when they are present on a page, are listed on the "Page information" page under "Lint errors".
If this list of pages is too long, you can increase the number of transclusions to 10 or some other number that gets the list to a reasonable size. I am expecting that the list will contain a few thousand pages. – Jonesey95 ( talk) 22:39, 30 June 2022 (UTC)
Pages like Red Banner of Labor, where the redirect target (case insensitive) contains the title, and is not in Redirects from short names. ― Qwerfjkl talk 22:24, 1 July 2022 (UTC)
Per
User talk:Qwerfjkl#Category:Austro-Hungarian films, I'd like to have all the edits by my bot that contain removed: [[Category:...]] ([2-7])
or if not possible, just the individual queries (or a template query, I suppose). From the edits I need the revision ID and the page name (I'm planning on mass-reverting all of the edits). ―
Qwerfjkl
talk 16:20, 5 July 2022 (UTC)
I would like to know all pages that have at least 80 color box template transclusions (not case sensitive). Not sure if this is relevant, but note that color boxes have multiple alternative template names, "{{Color box}}", "{{Colorbox}}", "{{Colour box}}", "{{Colourbox}}", "{{Color square}}", and "{{Legend3}}". TheGEICOgecko ( talk) 05:28, 4 July 2022 (UTC)
insource:"color box" insource:/(\{[Cc]olor ?box.+?){38}/
The first "insource" is needed because with the second one alone regex search times out. That one searches for any 38 repetitions of Colorbox/colorbox/Color box/... followed by any text until another Colorbox/... Just something to start with, don't expect all possible results. Ponor ( talk) 08:56, 6 July 2022 (UTC)
I'm trying to re-run a query from a few years ago and running into some issues.
AND ug_group = 'patroller'
to AND (ug_group = 'patroller' OR ug_group = 'sysop')
to fix this. –
Novem Linguae (
talk) 04:04, 7 July 2022 (UTC)
My filter rule is that the draft has only been edited by one person and that the current byte of the page is zero bytes (That is, the page is currently empty). Q 𝟤 𝟪 07:44, 14 July 2022 (UTC)
This is a query that tries to grab the most recent 1000 entries in the "Potential copyright violations log". You can view this log at Special:Log to see the kinds of entries the query is supposed to get. According to the PHP source code, 'pagetriage-copyvio' is the log type, but as you can see I am getting 0 results in my query. What'd I screw up? – Novem Linguae ( talk) 23:48, 18 July 2022 (UTC)
SHOW CREATE TABLE logging;
you'll see that rows with log_type = 'pagetriage-copyvio'
are filtered from the public replicas. Since they're visible without any needed user rights at
[2], and the log type is added only by an extension, it's reasonably likely they were accidentally left out of the whitelist rather than being deliberately omitted from it. You can try opening a phab task.Why are you looking at the pagetriage-copyvio log, anyway? It's singularly useless without the purported source of the infringement. —
Cryptic 01:04, 19 July 2022 (UTC)
I'm trying to get the most recent 5000 comments and scan them for some text using WHERE LIKE. Here's my attempt. However I imagine my LIMIT 5000 is applying to the first 5000 matches, not the first 5000 rows of the comment table, making my query slow. How do I get the first 5000 rows of the comment table, then scan for my text within those 5000 results? Subquery? – Novem Linguae ( talk) 06:28, 23 July 2022 (UTC)
I want to submit these files to QI so that I can review these pictures. Q 𝟤 𝟪 06:38, 21 July 2022 (UTC)
Hello!
I was hoping that I could be assisted with a Wiktionary query. I would like a query with all articles that meet two conditions: it should be a member of deepcat:"Non-lemma forms by language"
and Special:WhatLinksHere/PAGE should return 0 pages, it should hence be orphaned. I want in other words a list on all non-lemma words (regardless of language) that has 0 links to it. Grateful for answers.
Jonteemil (
talk) 23:12, 2 August 2022 (UTC)
Hello. Could someone please construct a query for all enwiki files in both category Self-published_work and category All_free_media, and have the two result columns be page.page_title and image.img_metadata? I tried in this query but I think I messed something up, because the resultset is bigger than I was expecting, and also because it is slow. Thanks. – Novem Linguae ( talk) 10:27, 6 August 2022 (UTC)
AND cl_type='file'
but it may make no difference: other pages shouldn't be in those categories. Of course, you'll need to join to page
and image
for the name and metadata.
Certes (
talk) 12:10, 6 August 2022 (UTC)
Hello. Does anyone know why, in my query of some files in the file namespace, most of the results have user_name = NULL? This suggests that almost every file in the 100,000 results has an unknown or logged out uploader. Also, deleting the "AND user_name NOT IN" part gets me the exact same number of results as having it in, and I was expecting it to eliminate some results. user_name is supposed to be the user name of the file uploader. Perhaps I messed something up. – Novem Linguae ( talk) 20:58, 9 August 2022 (UTC)
LEFT JOIN user ON user_id = actor_name
,
mw:Manual:Actor table. You want = actor_user. —
Cryptic 21:19, 9 August 2022 (UTC)Hello. Can I get help getting this query to exclude blocked users please? I tried adding the below and a couple variations (HAVING, wrapping the whole thing and making it a subquery) but couldn't quite crack it. Thank you.
JOIN user ON actor_user = user_id
LEFT JOIN ipblocks ON ipb_user = user_id
WHERE ipb_sitewide != 1
– Novem Linguae ( talk) 10:34, 12 August 2022 (UTC)
WHERE ipb_sitewide IS NULL OR ipb_sitewide != 1
. "
NULL != 1" isn't true. —
Cryptic 11:05, 12 August 2022 (UTC)Hi! I'd love a query written for the following conditions:
Thanks! 🐶 EpicPupper (he/him | talk) 21:01, 20 August 2022 (UTC)
For a major project report on NPP we need the following stats, if possible:
# | X | Description | Query |
---|---|---|---|
1 | ☑ | Number of new articles. | quarry:query/66788 |
2 | ☑ | Number of new articles created by users with < 90 days/500 edits. | quarry:query/66853 |
3 | ☑ | Number of new users with < 90 days/500 edits. who created articles and have been blocked for any reason. | quarry:query/66863 |
4 | ☑ | Number of new articles created by autopatrolled users. | quarry:query/66854 |
5 | ☑ | Number of new articles draftified. | quarry:query/66855 |
6 | ☐ | Number of new articles submitted to AfC | |
7 | ☑ | Number of drafts deleted including G13 | quarry:query/66789 |
8 | ☑ | Number of new articles deleted (by any deletion process). | quarry:query/66789 |
9 | ☑ | Number of NPPers who have not edited en.Wiki for 12 months. | quarry:query/66874 |
10 | ☑ | Number of NPPers who have not done any reviews for 12 months. | quarry:query/66875 |
Graphs are unavailable due to technical issues. There is more info on Phabricator and on MediaWiki.org. |
Stats curves on one graph. Sample period: 5 recent years (last 60 months)
I realise this might be a tall order, please let me know if any of these criteria are very awkward or impossible. Thanks, Kudpung กุดผึ้ง ( talk) 11:22, 20 August 2022 (UTC)
Hi Cryptic. By 'new article' I mean atricles in mainspace not including redirects or dabs. A draft is not a mainspace article, but an article created in, or moved to the draft namespace. I'll look at the quarries in the morning. Thanks. (FYI: Novem Linguae). Kudpung กุดผึ้ง ( talk) 17:22, 20 August 2022 (UTC)
Submitting using [[Wikipedia:AFCSW|AfC-submit-wizard]]
. Won't get them all, but should get an approximation. I checked 9 drafts at the top of the queue just now, 8 used that. Possible problems: 1) Will have duplicates (folks submitting multiple times). 2) Associated page may be deleted... not sure if this messes with the comment table? 3) Does the comment table try to recycle identical comments? If so that might mess with the count. –
Novem Linguae (
talk) 05:57, 25 August 2022 (UTC)
Hello! I'm back for another query for disambiguation pages with no WikiProject banner. This query should output articles that are:
Thanks! I tried tweaking the above query with this, but I'm horrible at SQL and it failed. 🐶 EpicPupper (he/him | talk) 00:30, 26 August 2022 (UTC)
Hi.
I am after a query to find out whether what % of new content do IP editors create. Wakelamp d[@-@]b ( talk) 13:00, 31 August 2022 (UTC)
This is an archive of past discussions. Do not edit the contents of this page. If you wish to start a new discussion or revive an old one, please do so on the current talk page. |
Archive 1 | Archive 2 | Archive 3 | Archive 4 |
Hi,
Per earlier discussion at Wikipedia talk:Naming conventions (geographic names)#Serbia village name format, we have a discrepancy in the formatting of disambiguation markers for villages in Serbia. Can someone please help produce a list of all articles in the main namespace that are underneath any of the categories of Category:Populated places in Serbia and match the regular expression /\(.*\)$/? We'd then be able to redact the list to make sure we don't have any false positives or negatives, and put this up for a mass move.
Likewise, if the query isn't too taxing on the available resources, we could do the analogous thing with articles underneath Category:Populated places in Bosnia and Herzegovina that have the same issue (it would be great to see what the impact of the mass move could be there).
TIA! -- Joy [shallot] ( talk) 07:21, 1 April 2021 (UTC)
.*\(.*\)$
. I put the category depth to 10, you can experiment with going deeper to make sure that is deep enough to include everything. I see a lot of non-villages mixed into the data... I think some of the deep sub-categories and/or their articles may need attention. For example,
Probus (emperor) is in the results because he is included in
Category:Sirmium, which was a Roman city that happened to be located in Serbia. Perhaps an additional category or a different RegEx would help to prune out these non-villages. –
Novem Linguae (
talk) 07:53, 1 April 2021 (UTC)
Chlod and I worked on this query, but both our attempts timed out.
Here's the code. Any suggestions for speeding this up? I think the revisions table has 1 billion rows, so it's quite a large table. Also, I read somewhere that WHERE, GROUP BY, and ORDER BY should be run on columns that are indexed/keyed for faster queries, but currently all 3 of those are being run on non-keyed columns. The
SQL optimizer tool is giving Query plan 1.1 is using filesort or a temporary table. This is usually an indication of an inefficient query. If you find your query is slow, try taking advantage of available indexes to avoid filesort.
–
Novem Linguae (
talk) 17:09, 2 April 2021 (UTC)
My approach was to check for pages in Category:All articles needing additional references. I tried Petscan, but couldn't find a filter for created date. I tried Quarry, and got it mostly working, but I got stuck on filtering by created date. I'm surprised there isn't a "created date" field in the `pages` table. Would appreciate any help. Even if it's to smack me for missing a simple solution :-) – Novem Linguae ( talk) 12:16, 12 April 2021 (UTC)
10+ years ago, as a first cutoff? For Wikipedia:Village pump (proposals)#Make page movers eligible to move move-protected pages – xeno talk 21:33, 26 April 2021 (UTC)
WHERE MAX(log_timestamp) < DATE_FORMAT(DATE_ADD(NOW(), INTERVAL -10 YEAR), "%Y%m%d%H%i%s")
, but it did not like WHERE MAX
, per
[1]. Again, if anyone sees a good solution, feel free to fork and post. Thanks. –
Novem Linguae (
talk) 22:41, 26 April 2021 (UTC)
Thomas Ranch is about to be deleted, having existed in Wikipedia for over fifteen years without a single inline reference, and without ever having an external link to an independent source. Grub Smith, though a bit longer, has been in a similar state for an equally long time. I happened to come across these two not while searching for suspect articles but while doing general cleanup around the given name Thomas and the surname Smith. How many more articles are out there in this condition? I propose that the best way to find out, if technically feasible, would be to generate a list of articles that have never had an inline ref tag over the course of their existence, sorted by age, and push through them from the oldest on forward. If someone has the Wiki-fu to generate such a list, please have at it. BD2412 T 02:42, 5 April 2021 (UTC)
I am thinking that the most immediate cases can be caught by an inquiry further narrowed to articles categorized in the Category:Companies category tree, or articles categorized in Category:Living people. The living people category will likely work well, but I found that digging down through the company category tree produces mostly dirt after just a few levels. I think you'll have more success grabbing articles that have a Company or company infobox (there are two of them, case sensitive). This set of articles is very clean. Sam at Megaputer ( talk) 23:50, 8 April 2021 (UTC)
In 2019, I
wrote a bot that added {{
unreferenced}}
to 10,000 articles. It could have added many more, but the community was dubious/conservative during an RfC and I had dial it back to the most extreme cases only. Thus any external link anywhere on the page disqualified it from being tagged, including a single external link in the external link section or in the infobox transcluded from a template. After it ran, there were zero complaints of bad tagging. It is capable of distinguishing sections and main body. It can create lists instead of tagging. It scours the entire 6+ million corpus looking for candidates. This is a non-trival bot even though it sounds easy-ish there are endless edge case exceptions - for example what counts as a dab vs. article? Not at all clear when you look at it. And so many external link templates to look for. It can't be done with a query it requires a purpose built application. The suggestion of looking for the worse of the worse is interesting, but I suspect the majority of articles tagged with {{
unreferenced}}
have always been that way. --
Green
C 02:33, 9 April 2021 (UTC)
In a discussion at the WP:Core Contest, I suggested a list of the most viewed Wikipedia articles that are start class or less would be beneficial for the competition, as presumably articles of interest to swathes of readers would be important ones to get right for our reputation. Obviously it may be useful to weed out the articles that are only prominent due to recent news stories (with a sharp spike in views e.g. Ron DeSantis ( views), in favour of those that have had more consistent page views through it's life on Wikipedia. I guess we could choose a certain standard deviation for page views to fall within from day to day? Note: Sometimes an article as different Class ratings from different Wikiprojects. Can you code in that the highest rating by the various Wikiprojects is the correct one? (Sometimes it is re-classed up but the other Wikiprojects don't catch up) I wouldn't worry about the Article Important class too much as many of these may not have been classed correctly.
-- Coin945 ( talk) 15:42, 29 May 2021 (UTC)
Hi, why isn't this working? I'm trying to generate a list of all indefinitely creation-protected articles at a sysop level. Anarchyte ( talk) 10:09, 4 June 2021 (UTC)
Thanks, Cryptic — that works perfectly. Anarchyte ( talk) 11:38, 4 June 2021 (UTC)
You are invited to join the discussion at Wikipedia talk:WikiProject Sweep § June 2021 follow-up. {{u| Sdkb}} talk 21:48, 4 June 2021 (UTC)
sub1
, such as filtering out disambig pages that are making up much of the output. –
SD0001 (
talk) 07:47, 20 June 2021 (UTC)
Hello!
The SQL code in c:Category talk:Animated PNG didn't work. Does anyone know how to fix it? The goal is to find all animated PNGs. Jonteemil ( talk) 18:17, 3 September 2021 (UTC)
Hello, this is a request I made on the talk page of the Database reports page, but it hasn't been answered over there. It is part of the overall proposal discussion I have started on the WikiProject Templates to create an Unused Templates Task Force to deal with the backlog of unused templates.
I've requested four reports per the original discussion and I'm going to relist them here:
I would like the report to run for at least two months as the task force is currently in an idea stage. When the reports are going to expire, if possible I would like to be notified of when it will happen. I need four reports from the Unused Templates database:
1) All unused templates that are not stubs or redirects.
2) All stub templates listed as unused. According to one of the users on the talk page discussion, there are either exactly or about 1,000 stub templates.
3) All redirects listed as unused. According to one of the users on the talk page discussion, there are either exactly or about 69,000 redirects.
4) Templates that were created and/or edited in the last year and the present year. -- WikiCleanerMan ( talk) 13:23, 28 September 2021 (UTC)
I'd like to refine this list (articles subcategorized under Category:United States to three levels that do not already have {{ Use mdy dates}} or {{ Use dmy dates}}) to remove all articles that are subcategorized under a different country. It'd also be nice to remove all articles that contain a date in DMY format outside of a reference. Would one or both of those be possible? {{u| Sdkb}} talk 19:08, 18 October 2021 (UTC)
Articles related to Canada may use either format with (as always) consistency within each article. We shouldn't be enforcing a standard date format on Canadian articles against the MOS, and against the fact that Canada doesn't exclusively use mdy dates. Joseph 2302 ( talk) 15:03, 19 October 2021 (UTC)
Your count for Linguistic Association of Canada and the United States is right (if "for" was a typo for "four"), but your route is off - Category:Education in Canada isn't a direct subcat of Category:Canada, and Category:Academic organizations based in Canada isn't a direct subcat of Category:Educational organizations based in Canada. It goes Category:Wikipedia categories named after countries (depth 0) > Category:Canada (depth 1) > Category:Organizations based in Canada (depth 2) > Category:Organizations based in Canada by subject (depth 3) > Category:Academic organizations based in Canada (depth 4).
What depth for Category:Military of the United States? (Just articles in that category is depth 0, articles in its direct subcat Category:American military aviation is depth 1, etc. to be sure we're talking about the same thing.)
The query excluding articles in depth 4 subcats of
Category:Wikipedia categories named after countries looks like it's going to die. Not all that surprising; depth 4 from
Category:United States is 8165 categories, depth 3 from
Category:Wikipedia categories named after countries is 39249, and depth 4 from it is 166700. —
Cryptic 21:24, 19 October 2021 (UTC) Of course it finished immediately after I gave up and saved the edit - 172888 results at depth 4 without doing anything with the Military subtree;
Linguistic Association of Canada and the United States and
Party for the Commonwealth of Canada correctly excluded, the French ship articles are still there. —
Cryptic 21:29, 19 October 2021 (UTC)
Hi. I'm looking for two related queries, one yielding a count, the other a handful of sample pages:
Extra credit: which red Draftspace pagename has the highest number of mainspace in-links, and how many? Thanks! Mathglot ( talk) 06:02, 25 October 2021 (UTC)
Basically, I'm looking for things like
{{cite journal ... |journal=Foobar |publisher=Foobar}}
{{cite magazine... |magazine=Foobar |publisher=Foobar}}
{{cite journal ... |work=Foobar |publisher=Foobar}}
{{cite web ... |website=Foobar |publisher=Foobar}}
{{cite news... |newspaper=Foobar |publisher=Foobar}}
{{citation ... |journal=Foobar |publisher=Foobar}}
E.g. if in a template you have any of [|journal=
, |magazine=
, |work=
, |website=
, |newspaper=
] that matches a |publisher=
in the same template, I'd want to know about it. It's OK if whitespace and casing differs. Thanks.
Headbomb {
t ·
c ·
p ·
b} 00:51, 17 October 2021 (UTC)
So last night my time I found this IP range with a rather interesting edit summary pattern, and thought I might try to use my limited MySQL knowledge to figure out how to use database queries to search for more instances of it. I made a working query for the RecentChanges table using query 24762 but my attempt to apply it to the whole revision table failed spectacularly (it seems to have hung (it's been running for about twelve hours now) but I can't figure out how to stop it). I originally tried searching for the phrase "%/info/en/?search=American_and_British_English_spelling_differences#/media/File:Defence_Defense_Labour_Labor_British_American_spelling_by_country.svg%" but this thread on mediawiki.org said that prefix searches like this are unbelievably slow, so I changed the query string to "China uses US spelling%" and simplified the query as far as I could without success. I can't figure out how to make use of the "revision_comment" alternative view in this case. Judging by the RecentChanges test and other independent searching of IP ranges that I've done, I'm not sure there'll be many more results, but if anyone can help me make a successful query for this info in the revision table, I'd appreciate it. Graham 87 08:31, 7 November 2021 (UTC)
Related to this idea, I'd like to create a bunch of redirects from e.g. Jane Lastname to Jane Q. Lastname. Would it be possible to compile a list of articles that meet the following criteria?
[word1] [capitalletter]. [word2]
[word1] [word2]
, nor any other articles with the title [word1] [anything] [word2]
Cheers, {{u| Sdkb}} talk 22:24, 21 November 2021 (UTC)
I
drafted a query for this but ran into some issues. Can you guys take a look and see if you can fix it? I'm trying to use the alias day
in some spots that it doesn't like, and I don't know how to fix it. Thanks. –
Novem Linguae (
talk) 02:18, 13 December 2021 (UTC)
I wrote a
query that gets an article's author. If I wanted to get this to get the authors of multiple articles in one query, what's the best way to refactor it? WHILE? Subquery? Here's my list of page IDs. Thank you. 66411662, 63051276, 67331300, 67494602, 67251604, 67475738, 67025524, 67282385, 67492243, 67505824, 68570713, 65754113, 68673796, 67481281, 68288689, 67485694, 68624634, 67062564, 67486327, 65912571, 67495969, 65558215, 67162967, 67504737, 66978220, 65952291, 67306801, 64208962, 67222236, 67365517, 68510913, 67480539, 66411751, 65228882, 67252944, 66476730, 68469744, 67008083, 66555751, 67282708, 67419043, 68693806
–
Novem Linguae (
talk) 16:57, 27 December 2021 (UTC)
I have a couple of stuck queries ( 61115, 58635) which still claim to be running hours beyond the max_statement_time I set. I'm aware that the tables are large, but I'm an experienced Quarry user who used to write SQL for a living, and the optimizer suggests that the queries use the expected indexes efficiently. The Stop button returns an HTML 500 response and seems to do nothing. Are these SQLs effectively dead, or is there some sort of regular clean-up process to release them? I'm reluctant to fork them or otherwise repeat my work in case it overloads the server. I remember that queries used to claim to be running after they had died for some reason (out of memory?) but in those cases it was possible to amend and re-run the query rather than it remaining permanently locked as "running". Certes ( talk) 00:29, 31 December 2021 (UTC)
Hello!
I would like a query of all files with MIME type image/x-bmp
and image-x-xcf
, respectively, across all wikis (all language versions of Wikipedia, Commons, Wikibooks, Mediawiki etc.). They can all be found at at
Special:MIMESearch/image/x-bmp and
Special:MIMESearch/image/x-xcf respectively on each wiki (or
Special:Search/file: local: filemime:bmp /
Special:Search/file: local: filemime:xcf). I would like the list to be rendered like this so a wikilink will work on any wiki:
w:File.NAME.bmp
fr:File.NAME.bmp
c:File:NAME.bmp
pt:v:File:NAME.bmp
etc.
Is that possible? To clarify I would like two queries, one per MIME type. Thanks! Jonteemil ( talk) 21:28, 9 January 2022 (UTC)
I need to check these article. Can anyone do this?-- 28 ( talk) 07:52, 10 January 2022 (UTC)
Can I have a list of all mainspace and Wikipedia namespace pages containing either spelling of the word "colo(u)r", or derivative words, regardless of redirect/disambiguation status, where there is no page with the other spelling? Animal lover 666 ( talk) 09:57, 13 February 2022 (UTC)
I'm trying to select (mainspace, non-redirect, non-dab) pages which have small gaps between their page_random value and the next-smallest page_random value (again, among mainspace non-redirect non-dab pages). I have a query here, which works when limited to a small subset of pages, but scales very badly as I relax the range of page_random values. I'd very much appreciate it if anyone could suggest any optimizations that would let me run it over the whole pages table (or at least larger chunks of it). Colin M ( talk) 18:35, 4 March 2022 (UTC)
This is purely for curiosity (or a possible Signpost feature), so take it up only if it's of interest. But I'd like to create a map of the world that shows how many good articles each country has per capita, and a similar one for FAs. The query for this would be articles in Category:Good articles (or Category:Featured articles) whose Wikidata item has country (P17) set to a given country. Those would then be adapted to shade an interactive map similar to Wugapodes' {{ Interactive COVID-19 maps}}. {{u| Sdkb}} talk 08:06, 17 March 2022 (UTC)
And here's for GAs with people, a larger data set:
Graphs are unavailable due to technical issues. There is more info on Phabricator and on MediaWiki.org. |
The winner is still Greenland, but Iceland's exceptional performance is looking less like a fluke. I wonder why that is (and Norway, too). And I expected India to do badly, but... geez. The only two visible countries to appear to have no GAs are Turkmenistan and Guinea-Bissau; I'll go offer rewards for those. {{u| Sdkb}} talk 19:17, 17 March 2022 (UTC)
Courtesy link (pushpin map): VPT Archive 196. Mathglot ( talk) 04:02, 2 April 2022 (UTC)
Hello, I'm interested in creating a list of orphans and articles with very few inlinks, which may be helpful as a To-do list for WikiProjects. For starters, and as an illustration and real-world example, I'm interested in a list of orphan articles that are categorized in WP:WikiProject Military history ("orphan" defined as an article with no in-links from other pages in mainspace; "in WikiProject X" means the associated talk page contains template {{ WikiProject X}}, or one of its redirects.) I was able to find this advanced search which is not quite the same thing but the "1,259" result may be a lower bound for the number of actual orphans.
Before generating any possibly long lists of articles, I'd like to first get a feeling for the scope of the problem. So, as a first db query can we do a summary query which returns eleven rows containing two values, where column one (let's call that 'nbr of inlinks') contains the integers 0 to 10, and column two ('article count') is the number of articles (ns=0) in WP:WikiProject Military history containing exactly that number of inlinks from other mainspace articles. So, e.g., the top data row contains a '0' on the left, and the number of orphans in WP:MILHIST on the right. The next row, a '1', and the count of MILHIST articles with exactly one (non-self) in-link from mainspace. And so on.
If the article count is not too huge, could we do a second query with one column containin the list of orphan articles in MILHIST as well? Please skip redirects, and I'm probably forgetting some other stuff, but anything that wouldn't be useful in a "ToDo" artile list for a wikiproject should be skipped as well. If you can request the article titles as a wikilink, that would be handy, and if it's not too difficult to add a second column containing a wikilinked Q-item from WikiData for those articles that have Wikidata links with at least one non-English Wikipedia link (i.e., >= 1 link excluding en-wiki and simple), that would be ideal. (Example: Olga Bogdanova is an orphan, but has d:Q26243976 attached to it, containing two non-English Wikipedia articles (ar, and ru); so in this case, col. 2 would be d:Q26243976. Thanks! Mathglot ( talk) 02:02, 2 April 2022 (UTC)
WITH RECURSIVE
in my
I-Spy book of SQL keywords. Analysis by link count is 0: 1540, 1: 5586, 2: 8008, 3: 9115, 4: 8892, 5: 7756, 6: 6817, 7: 6045, 8: 5156, 9: 4545, 10: 4104.
Certes (
talk) 23:32, 2 April 2022 (UTC)
This was about more than mere curiosity, and I hope it will lead to significant improvements in orphan backlog reduction. I've created a pilot at WikiProject Military history, and if you're interested in following along, see Wikipedia talk:WikiProject Military history#De-orphaning articles, and follow the link to the demo page. Mathglot ( talk) 01:01, 4 April 2022 (UTC)
As bots can't fix these. Thanks! 🐶 EpicPupper (he/him | talk) 21:50, 15 April 2022 (UTC)
AND pr_level IN ('sysop', 'interface-admin')
to the end of the first query, before the semicolon. I didn't mainly to prove that there aren't any interface-admin-protected ones, as opposed to a typo in the group name, or a missed rename like the recent 'oversight' to 'suppress'. (I'm fairly sure the current configuration disallows explicit protection to interface-admin anyway, and if we haven't changed the internal name from sysop to administrator yet we aren't likely ever to, but might as well be sure.) —
Cryptic 22:48, 15 April 2022 (UTC)Because of discussion at Help talk:Citation Style 1 § |location= without |publisher=, I wonder if there a way to get the number of articles that meet these criteria:
{{
cite book}}
template that has:
|location=<location>
AND
|publisher=
is empty OR|publisher=
is omitted{{
citation}}
template that:
|journal=<journal title>
OR|magazine=<magazine title>
OR|newspaper=<newspaper title>
OR|periodical=<periodical title>
OR|website=<website title>
OR|work=<work title>
|location=<location>
AND
|publisher=
is empty OR|publisher=
is omittedSure, we can modify Module:Citation/CS1 to add templates that match these criteria to a category but it might be better to know beforehand the magnitude of the problem if there is one.
All I want is the numbers. Thanks.
— Trappist the monk ( talk) 15:37, 23 April 2022 (UTC)
\s
does not match whitespace characters. For example,
this search does not find thousands of articles using {{
cite book}}
(with a space) but
this search does. Similarly, \w
is just an escaped w
... Rewriting your regex like this:
hastemplate:"cite book" insource:/\{ *cite book *(\| *(([^p]|p[^u]|pu[^b])[^\|\}]*|publisher *= *))*\| *location *= *[A-Za-z][^\|]*(\| *(([^p]|p[^u]|pu[^b])[^\|\}]*|publisher *= *))*\}\}/i
{{
cite book}}
with |location=
and without |publisher=
but, alas, all {{cite book}}
templates in
U.S. state, though returned by the search, have both |location=
and |publisher=
.Would it be possible to take the list here and filter it to only contain entries that have the wikitext "==Plot==" or "== Plot =="? {{u| Sdkb}} talk 02:23, 28 April 2022 (UTC)
Hi, I'd like to know if I can find out how many times a particular edit summary from a popular semi-automated tool has been used. Ohconfucious's MOS:NUMDATES script's former edit summary was: date formats per [[MOS:DATEFORMAT]] by [[WP:MOSNUMscript|script]]
(date formats per
MOS:DATEFORMAT by
script).
Dawnseeker2000 21:30, 8 May 2022 (UTC)
Hello, folks,
Is it possible to run a query on my own contributions and logs? This is not for a higher purpose, I am just curious about how many Teahouse invitations I've posted (I think it's in the tens of thousands) so I'd run a query to see how many of my edit summaries state "Welcome to Wikipedia: check out the Teahouse!" since that is the default edit summary when posting an invitation. I'm also curious about the pages that I've deleted, if I could run a query on the deletion summaries on Special:Log/Liz to see the proportion that were based on a particular speedy deletion criteria or whether they were deleted PRODs or because of an AfD decision. This query might be more complicated because sometimes multiple CSD criteria are listed in a deletion summary or the wording is not concise. But I don't think I've ever left a blank deletion summary so there would be some reason listed for every page deletion I've made.
If these tools are really intended to help manage the project, I understand that my request is purely personal. It's just that my edit count is growing and I know at a certain point in the future, Edit Count will no longer be available for me to see my how my edits breakdown so I thought I'd inquire about these two queries before my edit counts and log entries become too large for these to be practical requests.
Many thanks for any feedback you have, even if you tell me that these results are not possible to obtain. I'm constantly in awe of editors and admins who create bots or write scripts and make tools which make our editing here so much easier and simple. Your talents are appreciated! Thanks again. Liz Read! Talk! 20:24, 9 May 2022 (UTC)
Hey, I'm trying to list French sub talkpages 'À_faire' (like fr:Discussion:Dantès/À faire) that still exists while talkpage (like fr:Discussion:Dantès) OR main page (like fr:Dantès) doesn't exist.
I started with :
SELECT page_title
FROM page talkpage
WHERE talkpage.page_title LIKE '%/À_faire'
AND talkpage.page_namespace IN (1,5,11,15,101,119,711,829)
But, I'm not able to end this SQL query with NOT EXISTS
in order to check :
Maybe this is a starting point :
AND NOT EXISTS (
SELECT 1
FROM page mainpage
WHERE mainpage.page_namespace=talkpage.page_namespace-1
AND mainpage.page_title=talkpage.page_title)
But... it's going to list French sub talkpages 'À_faire' WHEN pages like fr:Dantès/À faire doesn't exist BUT that's not what I'm looking for. I need to check IF fr:Dantès OR fr:Talk:Dantès doesn't exist instead. Any clue?
Thank's -- LD ( talk) 22:10, 12 May 2022 (UTC)
AND NOT EXISTS (
SELECT 1
FROM page mainpage
WHERE mainpage.page_namespace IN (talkpage.page_namespace-1, talkpage.page_namespace)
AND mainpage.page_title=SUBSTRING(page_title,1,CHAR_LENGTH(page_title)-8))
CHAR_LENGTH('/À_faire')
. Did you also want to see pages where one of the two exists but not the other?
Certes (
talk) 22:50, 12 May 2022 (UTC)
AND (NOT EXISTS (SELECT 1 FROM page mainpage
WHERE mainpage.page_namespace = talkpage.page_namespace-1
AND mainpage.page_title=SUBSTRING(page_title,1,CHAR_LENGTH(page_title)-8))
OR NOT EXISTS (SELECT 1 FROM page mainpage
WHERE mainpage.page_namespace = talkpage.page_namespace
AND mainpage.page_title=SUBSTRING(page_title,1,CHAR_LENGTH(page_title)-8)))
Hello!
I'd like to find files whose extension don't match the MIME. I can find them using CirrusSearch, for example Special:Search/file: local: filemime:png -intitle:/\.png/i. You must however do this MIME by MIME - you can not do it all MIMEs simultaneously. It'd be nice to do a collective query of ALL files which don't match their MIME. You can find all possible file extensions and MIMEs on Special:MediaStatistics. Jonteemil ( talk) 19:58, 17 May 2022 (UTC)
Hi, I'm trying to re-run a bunch of queries on Quarry that I created a while back for looking for candidates to invite to New Page Patrol. For example: https://quarry.wmcloud.org/query/42657 among others. However, when I try to re-run them, it says "Bad database name". Could you indicate what needs to be changed to get them working again? — Insertcleverphrasehere( or here)( or here)( or here) 21:26, 26 May 2022 (UTC)
enwiki_p
in the "Enter the DB name here..." field between the description and the actual query
* Pppery *
it has begun... 21:29, 26 May 2022 (UTC)
Part 1: Need a query that gets a list of PageTriage/NPP unreviewed articles. This is tricky because I don't think there's a field anywhere that stores if an article is unreviewed. Rather, you have to JOIN logging
or JOIN pagetriage_log_table
and look for the absence of a record. Might also have to limit it by date, since PageTriage logs appear to be self-deleting after 12 months, so the absence of a log after 12 months = reviewed.
More info.
Part 2: Need a way to figure out if the unreviewed articles are species article or not. One of the best ways is to search the article text for "is a species of", but we can't search article text in SQL. Another way might be to check for any
biological sciences stub template. Another way might be to check for any article title that is in the format /[A-Z][a-z]+ [a-z]+/
(two words, first character is a capital letter, all other characters are lowercase) and I can do some manual filtering after that. –
Novem Linguae (
talk) 21:44, 26 May 2022 (UTC)
SELECT COUNT(*) FROM pagetriage_page JOIN page ON page_id=ptrp_page_id WHERE ptrp_reviewed=0 AND page_namespace=0 AND page_is_redirect=0;
at 14159. —
Cryptic 00:17, 27 May 2022 (UTC)
As suggested by
Sdkb
here, looking for a query that lists all articles in
Category:All Wikipedia level-5 vital articles but not in
Category:Articles with short description. Thank you! (I'm not
watching this page – please use {{
reply to|PerfectSoundWhatever}}
on reply) —
PerfectSoundWhatever (
t;
c) 15:01, 1 June 2022 (UTC)
Is it possible to obtain a list of recently updated talk pages efficiently, e.g. through SQL via page.page_latest? It could be a list of pages updated since a certain date, or a complete list of all pages with the date of last update, or possibly a list of all pages sorted by last update ranking. Per discussion Wikipedia:Bot_requests#Add_Reflist-talk_to_talk_pages_with_references. Green C 16:21, 3 June 2022 (UTC)
page_namespace in (1, 11)
shows just (article) Talk: and Template_talk:. I've excluded redirects: including them would speed the query up but produce lots of false positives.
Certes (
talk) 23:52, 3 June 2022 (UTC)
An edit request was made at
m:Module talk:Template translation where
Verdy p said that the language specified in the metadata of a page may not be the same with the last part of its title. For example,
m:User:Tangotango/Whatredirectshere/sv is expected to be in Swedish, but MediaWiki doesn't think so. I have tried writing a query myself using
this manual but page_lang
always return NULL
.
NguoiDungKhongDinhDanh 03:42, 15 June 2022 (UTC)
page_lang
means that page's language is similar to that of the project, then I'm looking for pages with at least one slash, without page_lang
set, where the part after the last slash isn't en
but matches the page_lang
of at least one other page. Honestly, I don't expect 1000, let alone 1.7 million.
NguoiDungKhongDinhDanh 07:42, 15 June 2022 (UTC)
page_lang
seems to be inconsistent with {{
PAGELANGUAGE}}. For example,
m:Main Page/es is clearly a Spanish translation (
expand). This request is about to be out of scope for this page so perhaps I'll need to ask at
WP:VPT. Thanks a lot anyway.
NguoiDungKhongDinhDanh 13:20, 15 June 2022 (UTC)I need these pages, but I don't know how to write a SQL query. Can anyone help me?-- Q28 ( talk) 14:21, 18 June 2022 (UTC)
I don't know if this is a query exactly, but I thought I would start here. What I am hoping to have generated is a table, a histogram of sorts, showing the number of articles that have short descriptions, put into bins by the length of the short description. Something like this:
Length (in characters) | Number of articles |
---|---|
N/A (no SD) | 2,000,000 |
1 | 10 |
2 | 100 |
3 | 265 |
... | ... |
99 | 2,204 |
100+ | 245 |
The purpose of this data request is to inform an ongoing conversation about the appropriate length of short descriptions.
Redirect pages should be excluded from the counts. The maximum bucket value should be "100+", since there should be only a hundred or so longer than that. One tricky bit is that some short descriptions are assigned by infobox templates, but I think that should be OK, since the descriptions are still listed under "Local description" on the Page information page for the article. According to "state of the project" data, there should be about 4.5 million articles with short descriptions and 1.6 million without them. The histogram does not need to be up-to-the-minute, so if it based on a database dump, that is fine. Please let me know if you have any questions. Thanks. – Jonesey95 ( talk) 03:19, 24 June 2022 (UTC)
For each AfD closed as redirect, how many times has the page been restored to an article and reverted to a redirect since then?
Example: WP:Articles for deletion/Children of This Earth, Special:PageHistory/Children of This Earth: 1 restore, 1 revert
Notes:
I think I may be able to figure this out if pointed in the right direction. Thanks! Flatscan ( talk) 04:21, 27 June 2022 (UTC)
I am looking for pages outside of Template namespace that have multiple transclusions, maybe 5 or more, and also have Linter errors. Linter errors, when they are present on a page, are listed on the "Page information" page under "Lint errors".
If this list of pages is too long, you can increase the number of transclusions to 10 or some other number that gets the list to a reasonable size. I am expecting that the list will contain a few thousand pages. – Jonesey95 ( talk) 22:39, 30 June 2022 (UTC)
Pages like Red Banner of Labor, where the redirect target (case insensitive) contains the title, and is not in Redirects from short names. ― Qwerfjkl talk 22:24, 1 July 2022 (UTC)
Per
User talk:Qwerfjkl#Category:Austro-Hungarian films, I'd like to have all the edits by my bot that contain removed: [[Category:...]] ([2-7])
or if not possible, just the individual queries (or a template query, I suppose). From the edits I need the revision ID and the page name (I'm planning on mass-reverting all of the edits). ―
Qwerfjkl
talk 16:20, 5 July 2022 (UTC)
I would like to know all pages that have at least 80 color box template transclusions (not case sensitive). Not sure if this is relevant, but note that color boxes have multiple alternative template names, "{{Color box}}", "{{Colorbox}}", "{{Colour box}}", "{{Colourbox}}", "{{Color square}}", and "{{Legend3}}". TheGEICOgecko ( talk) 05:28, 4 July 2022 (UTC)
insource:"color box" insource:/(\{[Cc]olor ?box.+?){38}/
The first "insource" is needed because with the second one alone regex search times out. That one searches for any 38 repetitions of Colorbox/colorbox/Color box/... followed by any text until another Colorbox/... Just something to start with, don't expect all possible results. Ponor ( talk) 08:56, 6 July 2022 (UTC)
I'm trying to re-run a query from a few years ago and running into some issues.
AND ug_group = 'patroller'
to AND (ug_group = 'patroller' OR ug_group = 'sysop')
to fix this. –
Novem Linguae (
talk) 04:04, 7 July 2022 (UTC)
My filter rule is that the draft has only been edited by one person and that the current byte of the page is zero bytes (That is, the page is currently empty). Q 𝟤 𝟪 07:44, 14 July 2022 (UTC)
This is a query that tries to grab the most recent 1000 entries in the "Potential copyright violations log". You can view this log at Special:Log to see the kinds of entries the query is supposed to get. According to the PHP source code, 'pagetriage-copyvio' is the log type, but as you can see I am getting 0 results in my query. What'd I screw up? – Novem Linguae ( talk) 23:48, 18 July 2022 (UTC)
SHOW CREATE TABLE logging;
you'll see that rows with log_type = 'pagetriage-copyvio'
are filtered from the public replicas. Since they're visible without any needed user rights at
[2], and the log type is added only by an extension, it's reasonably likely they were accidentally left out of the whitelist rather than being deliberately omitted from it. You can try opening a phab task.Why are you looking at the pagetriage-copyvio log, anyway? It's singularly useless without the purported source of the infringement. —
Cryptic 01:04, 19 July 2022 (UTC)
I'm trying to get the most recent 5000 comments and scan them for some text using WHERE LIKE. Here's my attempt. However I imagine my LIMIT 5000 is applying to the first 5000 matches, not the first 5000 rows of the comment table, making my query slow. How do I get the first 5000 rows of the comment table, then scan for my text within those 5000 results? Subquery? – Novem Linguae ( talk) 06:28, 23 July 2022 (UTC)
I want to submit these files to QI so that I can review these pictures. Q 𝟤 𝟪 06:38, 21 July 2022 (UTC)
Hello!
I was hoping that I could be assisted with a Wiktionary query. I would like a query with all articles that meet two conditions: it should be a member of deepcat:"Non-lemma forms by language"
and Special:WhatLinksHere/PAGE should return 0 pages, it should hence be orphaned. I want in other words a list on all non-lemma words (regardless of language) that has 0 links to it. Grateful for answers.
Jonteemil (
talk) 23:12, 2 August 2022 (UTC)
Hello. Could someone please construct a query for all enwiki files in both category Self-published_work and category All_free_media, and have the two result columns be page.page_title and image.img_metadata? I tried in this query but I think I messed something up, because the resultset is bigger than I was expecting, and also because it is slow. Thanks. – Novem Linguae ( talk) 10:27, 6 August 2022 (UTC)
AND cl_type='file'
but it may make no difference: other pages shouldn't be in those categories. Of course, you'll need to join to page
and image
for the name and metadata.
Certes (
talk) 12:10, 6 August 2022 (UTC)
Hello. Does anyone know why, in my query of some files in the file namespace, most of the results have user_name = NULL? This suggests that almost every file in the 100,000 results has an unknown or logged out uploader. Also, deleting the "AND user_name NOT IN" part gets me the exact same number of results as having it in, and I was expecting it to eliminate some results. user_name is supposed to be the user name of the file uploader. Perhaps I messed something up. – Novem Linguae ( talk) 20:58, 9 August 2022 (UTC)
LEFT JOIN user ON user_id = actor_name
,
mw:Manual:Actor table. You want = actor_user. —
Cryptic 21:19, 9 August 2022 (UTC)Hello. Can I get help getting this query to exclude blocked users please? I tried adding the below and a couple variations (HAVING, wrapping the whole thing and making it a subquery) but couldn't quite crack it. Thank you.
JOIN user ON actor_user = user_id
LEFT JOIN ipblocks ON ipb_user = user_id
WHERE ipb_sitewide != 1
– Novem Linguae ( talk) 10:34, 12 August 2022 (UTC)
WHERE ipb_sitewide IS NULL OR ipb_sitewide != 1
. "
NULL != 1" isn't true. —
Cryptic 11:05, 12 August 2022 (UTC)Hi! I'd love a query written for the following conditions:
Thanks! 🐶 EpicPupper (he/him | talk) 21:01, 20 August 2022 (UTC)
For a major project report on NPP we need the following stats, if possible:
# | X | Description | Query |
---|---|---|---|
1 | ☑ | Number of new articles. | quarry:query/66788 |
2 | ☑ | Number of new articles created by users with < 90 days/500 edits. | quarry:query/66853 |
3 | ☑ | Number of new users with < 90 days/500 edits. who created articles and have been blocked for any reason. | quarry:query/66863 |
4 | ☑ | Number of new articles created by autopatrolled users. | quarry:query/66854 |
5 | ☑ | Number of new articles draftified. | quarry:query/66855 |
6 | ☐ | Number of new articles submitted to AfC | |
7 | ☑ | Number of drafts deleted including G13 | quarry:query/66789 |
8 | ☑ | Number of new articles deleted (by any deletion process). | quarry:query/66789 |
9 | ☑ | Number of NPPers who have not edited en.Wiki for 12 months. | quarry:query/66874 |
10 | ☑ | Number of NPPers who have not done any reviews for 12 months. | quarry:query/66875 |
Graphs are unavailable due to technical issues. There is more info on Phabricator and on MediaWiki.org. |
Stats curves on one graph. Sample period: 5 recent years (last 60 months)
I realise this might be a tall order, please let me know if any of these criteria are very awkward or impossible. Thanks, Kudpung กุดผึ้ง ( talk) 11:22, 20 August 2022 (UTC)
Hi Cryptic. By 'new article' I mean atricles in mainspace not including redirects or dabs. A draft is not a mainspace article, but an article created in, or moved to the draft namespace. I'll look at the quarries in the morning. Thanks. (FYI: Novem Linguae). Kudpung กุดผึ้ง ( talk) 17:22, 20 August 2022 (UTC)
Submitting using [[Wikipedia:AFCSW|AfC-submit-wizard]]
. Won't get them all, but should get an approximation. I checked 9 drafts at the top of the queue just now, 8 used that. Possible problems: 1) Will have duplicates (folks submitting multiple times). 2) Associated page may be deleted... not sure if this messes with the comment table? 3) Does the comment table try to recycle identical comments? If so that might mess with the count. –
Novem Linguae (
talk) 05:57, 25 August 2022 (UTC)
Hello! I'm back for another query for disambiguation pages with no WikiProject banner. This query should output articles that are:
Thanks! I tried tweaking the above query with this, but I'm horrible at SQL and it failed. 🐶 EpicPupper (he/him | talk) 00:30, 26 August 2022 (UTC)
Hi.
I am after a query to find out whether what % of new content do IP editors create. Wakelamp d[@-@]b ( talk) 13:00, 31 August 2022 (UTC)