![]() | This is an archive of past discussions. Do not edit the contents of this page. If you wish to start a new discussion or revive an old one, please do so on the current talk page. |
Archive 35 | Archive 36 | Archive 37 | Archive 38 | Archive 39 | Archive 40 | → | Archive 45 |
Hi Cyberpower678, it looks as if memoire-du-cyclisme.net has been dead since 2012. Can all links from that year be rescued? Would be great if your bot could run through and archive them from that year. This site is heavily used in cycling-related articles. In my case it was 1990 Tour de France. I'm not sure why this hasn't been picked up sooner. Thanks. Bald Boris 15:21, 5 September 2016 (UTC)
Am I missing something? I'm only asking for help because this site is used to cite the most important info, in my case, the Tour de Frances, as I'm sorting all of them out. Can you let me know if this can be done with the InternetArchiveBot? Otherwise I guess I'll have to individually archive them? Bald Boris 16:40, 6 September 2016 (UTC)
{{
dead link}}
to them will it fetch an archive before the access date?
Bald
Boris
21:24, 6 September 2016 (UTC)The bot signature HTML is broken and leaving dirty wiki-markup on every page it appears on, wrecking the display of wiki-markup for all subsequent content on a talk page. The signature fails to close the <sup> and <span> tags. This is extremely discourteous. Please fix this ASAP, thanks. — btphelps ( talk to me) ( what I've done) 07:11, 7 September 2016 (UTC)
This bot is making two "weird" edits, that I don't think are inline with what it is supposed to do, or are actually correct. Both at ...Ye Who Enter Here and 0-8-4 it is adding extra url "crap" as I call it to the webcitation links, which is not needed. Is this correct? Is it attempting to fix another issue I'm not aware of and subsequently adding this text? Thanks for the input. - Favre1fan93 ( talk) 22:28, 8 September 2016 (UTC)
Hi Cyberpower678, I noticed that cyberbot II made this edit, marking the dead URL as unfit. The archive URL does look like a wild 90's site, but otherwise doesn't seem (to me) to meet the "unfit" definition at Category:CS1 maint: Unfit url. Is Cyberbot advising use of a different site? Is there a list that Cyberbot is checking? Clarification or advice would be appreciated! - Paul2520 ( talk) 02:42, 9 September 2016 (UTC)
this edit is pretty bad. the bot removes the back half of an HTML comment, but leaves the front half. can you make sure this was the only one? Frietjes ( talk) 20:44, 8 September 2016 (UTC)
Hello. Please explain how my request here triggered your bot. Was it an error? Thanks. X4n6 ( talk) 03:43, 10 September 2016 (UTC)
There are cases of protocol-relative URLs eg. |archiveurl=//web.archive.org/
in
Ewa Kopacz. At
one time consensus was to use PRURL for Wayback but the last RfC said to convert to https.
[2] This can easily be fixed with AWB, but it might mess up the CBDB if the changes are not sent in via SQL. How does the bot handle PRURL's in the CBDB, are they converted to https, retained, skipped? Do you want to add a feature to IABot to do the convert? Should I add it to wayback medic (though it's now 25% done, I could go back over those with only the prurl feature enabled) --
Green
C
14:45, 6 September 2016 (UTC)
Template:Adminstats/!dea4u has been
nominated for deletion. You are invited to comment on the discussion at
the template's entry on the Templates for discussion page. ~
Rob13
Talk
05:56, 11 September 2016 (UTC)
Hi Cyberpower, just a small thing: at RFPP, Cyberbot seems to think that there is 1 (more?) pending request than there actually is. This might be a bug. Best, Airplaneman ✈ 10:31, 11 September 2016 (UTC)
https://tools.wmflabs.org/xtools-ec/?user=Kudpung&project=en.wikipedia.org is not working. I need some of this info because it required by Arbcom. I have notified the other developers. -- Kudpung กุดผึ้ง ( talk) 06:34, 11 September 2016 (UTC)
The diffs are looking very good. It sometimes/often has trouble with deadurl=no. [3] [4] [5] [6].. should be yes or not bot unknown. -- Green C 14:49, 8 September 2016 (UTC)
I'm finding cases where "/web/" matters. Example:
The #2 returns a header "HTTP/1.1 400 Invalid URI: noSlash". The API returned #2 (wrong). So this will be a new task of WM to remove "/web/" if the URL isn't working but the non-"/web" version is working. It's already coded going forward.
It also means automatically reformatting existing URLs to "/web" might be a problem. WM2 has been doing this and unaware deleting links with this problem since the new formatted URL didn't verify. I don't know how many probably a couple hundred. I can figure out which ones if IABot will accept an SQL update with new archive URLs that don't contain "/web/". -- Green C 15:08, 11 September 2016 (UTC)
https://en.wikipedia.org/?title=Malta&oldid=739116927
The above edit by your bot cut out 76kb of article content. Something is amiss. Rhialto ( talk) 22:00, 12 September 2016 (UTC)
Hey Cyberpower, it looks like the user talk page didn't get moved when renaming this user. clpo13( talk) 16:13, 13 September 2016 (UTC)
Appreciated! ronaz Talk! 17:04, 13 September 2016 (UTC)
Inclusion of :80 is breaking links. Examples:
The :80 is being added by the API. I believe it may be there as convenience information for the calling application and not reflect the actual Wayback link. -- Green C 01:29, 12 September 2016 (UTC)
It looks like the problem is more complicated. If port is 80 and the source URL protocol is https then it doesn't work because https is port 443. Example from John Major:
If the API is returning an https URL with port 80 that's a problem. However I can't replicate, the API doesn't convert the source http URL to https. -- Green C 19:39, 12 September 2016 (UTC)
Could it be caused by a greedy regex when changing https for the archive.org portion it bleeds into the rest of the URL? -- Green C 18:00, 13 September 2016 (UTC)
https://archive.org.../http://source.com...
. It will convert
http://source.com to
https://source.com because
https://archive.org is already set to https. --
Green
C
20:23, 13 September 2016 (UTC)
Hi Cyberpower678. In this edit, InternetArchiveBot added "deadurl=no" to three references, but in each case, the original URL produces a "couldn't find this page on the site" type message. Is that correct/expected behaviour? Regards. DH85868993 ( talk) 06:55, 14 September 2016 (UTC)
It doesn't look like it's noticed memoire-du-cyclisme.net is dead with 1938 Tour de France [12] and 1930 Tour de France [13]. Bald Boris 19:42, 13 September 2016 (UTC)
Hello. It says on the FAQ that such instances should be notified to you to reset the status, so here I am. With this edit earlier today, the bot added an archive link for two pages from the rsssf.com website and set the deadurl param to yes, but they're both live now. cheers, Struway2 ( talk) 13:48, 14 September 2016 (UTC)
I received a warning about this edit, where I removed an AFD notice from a page. I removed the notice because draft space articles belong in MFD. So the removal was valid. Cheers and thanks for all the good work this bot does.-- Adam in MO Talk 17:06, 17 September 2016 (UTC)
Hi Cyberpower678, I'm following up again about {{ X13}} .. {{ X20}} and their talk pages. It seems that the bot is still unaware of these the relatively new template sandboxes, and I'm thinking it makes sense for the bot to reset them at some point. — Andy W. ( talk · ctb) 00:16, 18 September 2016 (UTC)
Note that this bot is blocked to access webpages from the Minor Planet Center. E.g. this page.. see revert. I have already posted this fact before on this very page here. For the last few weeks, I have spent hours reverting these unhelpful bots edits and flagged the corresponding post on the talk-page with FAILED. This has to stop. Rfassbind – talk 20:46, 15 September 2016 (UTC)
is being forwarded to:
it is detrimental to label it as dead link and add an unhelpful, completely outdated archive link. The HTTP Status Code for the erroneously considered dead URI is a 302 found. Rfassbind – talk 00:30, 18 September 2016 (UTC)
Hello friend! It looks like Cyberbot I has taken a break from clerking WP:RFPP? Perhaps you could kick it back into order? Best — MusikAnimal talk 04:15, 23 September 2016 (UTC)
I undid this edit by User:InternetArchiveBot since both of the links it marked as dead (i.e. http://www.physics.sfasu.edu/astro/asteroids/sizemagnitude.html and http://www.gps.caltech.edu/~mbrown/dps.html) appear to be live links, not dead links, as they both function fine for me. Just in case this is indicative of some kind of bug in InternetArchiveBot and not some sort of transitory internet connection issue for the bot, I'm leaving this bug report. — RP88 ( talk) 01:33, 25 September 2016 (UTC)
I received a warning about this edit, where I removed an AFD notice from a page. I removed the notice because draft space articles belong in MFD. So the removal was valid. Cheers and thanks for all the good work this bot does.-- Adam in MO Talk 17:06, 17 September 2016 (UTC)
Hi Cyberpower678, I'm following up again about {{ X13}} .. {{ X20}} and their talk pages. It seems that the bot is still unaware of these the relatively new template sandboxes, and I'm thinking it makes sense for the bot to reset them at some point. — Andy W. ( talk · ctb) 00:16, 18 September 2016 (UTC)
Note that this bot is blocked to access webpages from the Minor Planet Center. E.g. this page.. see revert. I have already posted this fact before on this very page here. For the last few weeks, I have spent hours reverting these unhelpful bots edits and flagged the corresponding post on the talk-page with FAILED. This has to stop. Rfassbind – talk 20:46, 15 September 2016 (UTC)
is being forwarded to:
it is detrimental to label it as dead link and add an unhelpful, completely outdated archive link. The HTTP Status Code for the erroneously considered dead URI is a 302 found. Rfassbind – talk 00:30, 18 September 2016 (UTC)
I assume, youbot doesn't go trough all websites, right, but only on those, which are in your database? If so (and this post is useful), you can add *.london2012.com to list. If this wasn't useful post of mine (and bot archives really everything it can), say so :) --
Edgars2007 (
talk/
contribs)
19:10, 24 September 2016 (UTC)
In the edit your bot made ( diff link) three Alexa.com links got tagged as dead but where working when I checked. I should also let you know (if you don't already) that the Wayback Machine has had intermittent failures lately. Your bot should avoid substituting "https" for "http" within URLs or adding the protocol number ":80" (the Wayback Machine has done this) as another bot GreenC bot ( talk · contribs) cleans up such links as it did with the follow up edit. Regards. – Allen4 names ( contributions) 11:42, 25 September 2016 (UTC)
When IABot isn't provided the accessdate, what happens? Also does the IABot add archive urls to urls that aren't necessarily dead? -- Tim 1357 talk| poke 05:57, 25 September 2016 (UTC)
I was unhappy with this edit by the InternetArchiveBot.
I have reverted the edit, and tagged the article with a {{bots|deny=InternetArchiveBot}} template Hawkeye7 ( talk) 22:18, 26 September 2016 (UTC)
{{
cbignore}}
.—
cyberpower
Chat:Offline
22:32, 26 September 2016 (UTC)
{{
cbignore}}
documentation was unclear as to whether it would work in this case.Hawkeye7 ( talk) 23:03, 26 September 2016 (UTC)
{{
cbignore}}
works on anything the bot touches. It's parsing engine is quite intelligent. It's designed to mimic human parsing as best as possible.|dead-url=bot: unknown
is not intended for use by regular editors, it is only vaguely documented. It functions the same way as |dead-url=unfit
, used previously by this bot, except that it is semantically different and has its own maintenance category:
Category:CS1 maint: BOT: original-url status unknown. |dead-url=unfit
(or |dead-url=usurped
) is intended to be used by editors who find an original url that may once have supported an article's text but now has been taken over for the purposes of advertising, porn, malicious scripts, or whatever. When these kinds of urls are found, |dead-url=unfit
causes
Module:Citation/CS1 to hide that url while at the same time retaining it for historical purposes. This bot was using |dead-url=unfit
which has contaminated
Category:CS1 maint: Unfit url so now we don't know which of the 9k-ish original urls in that category are unfit and which the bot knows to be live but doesn't know if they support their respective articles.|df=
is documented for all cs1|2 templates that use {{
csdoc}}
. See
Template:Cite web#Date.There's a problem with this edit by the bot: For the access date, it inserted the date the archive was captured, not the (presumably) current date when the bot accessed the page. This date precedes the founding of Wikipedia, which correctly caused it to be flagged as an error. -- Floatjon ( talk) 15:58, 29 September 2016 (UTC)
I can't believe no one hasn't already asked you this but fer chrissakes adjust the bot so it stops spamming talk pages. The edit summary in the pages' histories is more than enough notice that the external links have been modified. Such notification is in fact precisely what they are there for. The talk page is for actual talk about the improvement of the article which "double checking bot edits" doesn't actually fall under. The only pages where anyone ever follows the bot's instructions, verifies the link, and makes a mark are well-curated pages where they would have checked the new links anyway.
At minimum—if there is some policy that requires such notifications—make them as minimal as possible, as a single text line without images: I have just added archive links to one external link on X. Please take a moment to review my edit. If necessary, add after the link to keep me from modifying it. Alternatively, you can add to keep me off the page altogether.
. Right now, it's an eyesore. Nothing else being written is important, creating
a category with 180,000+ entries neither you nor anyone else is whittling down is not helping anyone, and it is a needless piece of makework to require other people delete them by hand or craft other bots to follow around cleaning up after you. —
LlywelynII
02:42, 27 September 2016 (UTC)
|checked=failed
, thinking that it might be useful feedback. I think myself a little unusual in going to such trouble. On even my most active articles, I have little competition to complete this task. I neglect to change archived messages on pages I'm new to, even if I check the link, since changing archived material is generally discouraged. The failed links that go undetected should be either relisted or reworked with time, especially since the bot seems to work on automatic pilot in checking for failed links.
Dhtwiki (
talk)
23:38, 29 September 2016 (UTC)
on the 2 Wayback sources you added here. a couple of hours ago (I think). Could you please take a look? Thanks, -- Hordaland ( talk) 04:32, 1 October 2016 (UTC)
Thinking about {{
official website}}
(aka {{
official}}
), see this
for example. This is a complex template as it gets link data from Wikidata and compares it with the source link, and if they don't match it considers it an error and adds to a tracking category. In this example that's what happened. I can see three possible solutions:
{{
wayback}}
to the end eg: *{{Official|example.com}} {{webarchive |url=https://web.archive.org/web/*/example.com |date=* }}
*{{webarchive |url=https://web.archive.org/web/*/example.com |date=* }}. In this case it would also have to transfer any
|name=
argument to wayback's |title=
.Each method has pros and cons. Official is a widely used template with 162192 transclusion, more than double the number of wayback templates. Probably the least disruptive would be #2 because it doesn't touch the original official, though it presents problems if the archive isn't templated like wayback or webcite. #3 would be the best performance since it reduces number of templates. #1 would be smooth but increases complexity of that template code and unclear there is consensus for it. And it will be up to you what is better solution for the bot. -- Green C 13:32, 30 September 2016 (UTC)
This nested a dead links template inside another template creating an error on render. -- Green C 13:08, 30 September 2016 (UTC)
{{
official}}
template, because it is neither an archive template or a cite template, nor does it share any cite template functionality. So it just sees the link inside it.—
cyberpower
Chat:Online
12:21, 1 October 2016 (UTC)
Hi! In this edit, InternetArchiveBot added "deadurl=no" and an empty "df" parameter. First of all, the original URL was dead when I tested it ("server not found"), so the parameter should have been "deadurl=yes". Second, why the empty date-format parameter? Is that supposed to be filled in later? Thanks. — Gorthian ( talk) 00:52, 2 October 2016 (UTC)
I just noticed on WP:RFPP that if a request is not [ signed Cyberbot I will make an edit saying it can't be parsed. Scroll down and there is a error message caused by the lack of signature in the original request. This causes the bot to return and tell itself that the previous bot message can't be [ parsed. Don't know if you have some way to suppress the error message in the first bot comment or stop the second. Cheers. CambridgeBayWeather, Uqaqtuq (talk), Sunasuttuq 09:25, 2 October 2016 (UTC)
![]() | This is an archive of past discussions. Do not edit the contents of this page. If you wish to start a new discussion or revive an old one, please do so on the current talk page. |
Archive 35 | Archive 36 | Archive 37 | Archive 38 | Archive 39 | Archive 40 | → | Archive 45 |
Hi Cyberpower678, it looks as if memoire-du-cyclisme.net has been dead since 2012. Can all links from that year be rescued? Would be great if your bot could run through and archive them from that year. This site is heavily used in cycling-related articles. In my case it was 1990 Tour de France. I'm not sure why this hasn't been picked up sooner. Thanks. Bald Boris 15:21, 5 September 2016 (UTC)
Am I missing something? I'm only asking for help because this site is used to cite the most important info, in my case, the Tour de Frances, as I'm sorting all of them out. Can you let me know if this can be done with the InternetArchiveBot? Otherwise I guess I'll have to individually archive them? Bald Boris 16:40, 6 September 2016 (UTC)
{{
dead link}}
to them will it fetch an archive before the access date?
Bald
Boris
21:24, 6 September 2016 (UTC)The bot signature HTML is broken and leaving dirty wiki-markup on every page it appears on, wrecking the display of wiki-markup for all subsequent content on a talk page. The signature fails to close the <sup> and <span> tags. This is extremely discourteous. Please fix this ASAP, thanks. — btphelps ( talk to me) ( what I've done) 07:11, 7 September 2016 (UTC)
This bot is making two "weird" edits, that I don't think are inline with what it is supposed to do, or are actually correct. Both at ...Ye Who Enter Here and 0-8-4 it is adding extra url "crap" as I call it to the webcitation links, which is not needed. Is this correct? Is it attempting to fix another issue I'm not aware of and subsequently adding this text? Thanks for the input. - Favre1fan93 ( talk) 22:28, 8 September 2016 (UTC)
Hi Cyberpower678, I noticed that cyberbot II made this edit, marking the dead URL as unfit. The archive URL does look like a wild 90's site, but otherwise doesn't seem (to me) to meet the "unfit" definition at Category:CS1 maint: Unfit url. Is Cyberbot advising use of a different site? Is there a list that Cyberbot is checking? Clarification or advice would be appreciated! - Paul2520 ( talk) 02:42, 9 September 2016 (UTC)
this edit is pretty bad. the bot removes the back half of an HTML comment, but leaves the front half. can you make sure this was the only one? Frietjes ( talk) 20:44, 8 September 2016 (UTC)
Hello. Please explain how my request here triggered your bot. Was it an error? Thanks. X4n6 ( talk) 03:43, 10 September 2016 (UTC)
There are cases of protocol-relative URLs eg. |archiveurl=//web.archive.org/
in
Ewa Kopacz. At
one time consensus was to use PRURL for Wayback but the last RfC said to convert to https.
[2] This can easily be fixed with AWB, but it might mess up the CBDB if the changes are not sent in via SQL. How does the bot handle PRURL's in the CBDB, are they converted to https, retained, skipped? Do you want to add a feature to IABot to do the convert? Should I add it to wayback medic (though it's now 25% done, I could go back over those with only the prurl feature enabled) --
Green
C
14:45, 6 September 2016 (UTC)
Template:Adminstats/!dea4u has been
nominated for deletion. You are invited to comment on the discussion at
the template's entry on the Templates for discussion page. ~
Rob13
Talk
05:56, 11 September 2016 (UTC)
Hi Cyberpower, just a small thing: at RFPP, Cyberbot seems to think that there is 1 (more?) pending request than there actually is. This might be a bug. Best, Airplaneman ✈ 10:31, 11 September 2016 (UTC)
https://tools.wmflabs.org/xtools-ec/?user=Kudpung&project=en.wikipedia.org is not working. I need some of this info because it required by Arbcom. I have notified the other developers. -- Kudpung กุดผึ้ง ( talk) 06:34, 11 September 2016 (UTC)
The diffs are looking very good. It sometimes/often has trouble with deadurl=no. [3] [4] [5] [6].. should be yes or not bot unknown. -- Green C 14:49, 8 September 2016 (UTC)
I'm finding cases where "/web/" matters. Example:
The #2 returns a header "HTTP/1.1 400 Invalid URI: noSlash". The API returned #2 (wrong). So this will be a new task of WM to remove "/web/" if the URL isn't working but the non-"/web" version is working. It's already coded going forward.
It also means automatically reformatting existing URLs to "/web" might be a problem. WM2 has been doing this and unaware deleting links with this problem since the new formatted URL didn't verify. I don't know how many probably a couple hundred. I can figure out which ones if IABot will accept an SQL update with new archive URLs that don't contain "/web/". -- Green C 15:08, 11 September 2016 (UTC)
https://en.wikipedia.org/?title=Malta&oldid=739116927
The above edit by your bot cut out 76kb of article content. Something is amiss. Rhialto ( talk) 22:00, 12 September 2016 (UTC)
Hey Cyberpower, it looks like the user talk page didn't get moved when renaming this user. clpo13( talk) 16:13, 13 September 2016 (UTC)
Appreciated! ronaz Talk! 17:04, 13 September 2016 (UTC)
Inclusion of :80 is breaking links. Examples:
The :80 is being added by the API. I believe it may be there as convenience information for the calling application and not reflect the actual Wayback link. -- Green C 01:29, 12 September 2016 (UTC)
It looks like the problem is more complicated. If port is 80 and the source URL protocol is https then it doesn't work because https is port 443. Example from John Major:
If the API is returning an https URL with port 80 that's a problem. However I can't replicate, the API doesn't convert the source http URL to https. -- Green C 19:39, 12 September 2016 (UTC)
Could it be caused by a greedy regex when changing https for the archive.org portion it bleeds into the rest of the URL? -- Green C 18:00, 13 September 2016 (UTC)
https://archive.org.../http://source.com...
. It will convert
http://source.com to
https://source.com because
https://archive.org is already set to https. --
Green
C
20:23, 13 September 2016 (UTC)
Hi Cyberpower678. In this edit, InternetArchiveBot added "deadurl=no" to three references, but in each case, the original URL produces a "couldn't find this page on the site" type message. Is that correct/expected behaviour? Regards. DH85868993 ( talk) 06:55, 14 September 2016 (UTC)
It doesn't look like it's noticed memoire-du-cyclisme.net is dead with 1938 Tour de France [12] and 1930 Tour de France [13]. Bald Boris 19:42, 13 September 2016 (UTC)
Hello. It says on the FAQ that such instances should be notified to you to reset the status, so here I am. With this edit earlier today, the bot added an archive link for two pages from the rsssf.com website and set the deadurl param to yes, but they're both live now. cheers, Struway2 ( talk) 13:48, 14 September 2016 (UTC)
I received a warning about this edit, where I removed an AFD notice from a page. I removed the notice because draft space articles belong in MFD. So the removal was valid. Cheers and thanks for all the good work this bot does.-- Adam in MO Talk 17:06, 17 September 2016 (UTC)
Hi Cyberpower678, I'm following up again about {{ X13}} .. {{ X20}} and their talk pages. It seems that the bot is still unaware of these the relatively new template sandboxes, and I'm thinking it makes sense for the bot to reset them at some point. — Andy W. ( talk · ctb) 00:16, 18 September 2016 (UTC)
Note that this bot is blocked to access webpages from the Minor Planet Center. E.g. this page.. see revert. I have already posted this fact before on this very page here. For the last few weeks, I have spent hours reverting these unhelpful bots edits and flagged the corresponding post on the talk-page with FAILED. This has to stop. Rfassbind – talk 20:46, 15 September 2016 (UTC)
is being forwarded to:
it is detrimental to label it as dead link and add an unhelpful, completely outdated archive link. The HTTP Status Code for the erroneously considered dead URI is a 302 found. Rfassbind – talk 00:30, 18 September 2016 (UTC)
Hello friend! It looks like Cyberbot I has taken a break from clerking WP:RFPP? Perhaps you could kick it back into order? Best — MusikAnimal talk 04:15, 23 September 2016 (UTC)
I undid this edit by User:InternetArchiveBot since both of the links it marked as dead (i.e. http://www.physics.sfasu.edu/astro/asteroids/sizemagnitude.html and http://www.gps.caltech.edu/~mbrown/dps.html) appear to be live links, not dead links, as they both function fine for me. Just in case this is indicative of some kind of bug in InternetArchiveBot and not some sort of transitory internet connection issue for the bot, I'm leaving this bug report. — RP88 ( talk) 01:33, 25 September 2016 (UTC)
I received a warning about this edit, where I removed an AFD notice from a page. I removed the notice because draft space articles belong in MFD. So the removal was valid. Cheers and thanks for all the good work this bot does.-- Adam in MO Talk 17:06, 17 September 2016 (UTC)
Hi Cyberpower678, I'm following up again about {{ X13}} .. {{ X20}} and their talk pages. It seems that the bot is still unaware of these the relatively new template sandboxes, and I'm thinking it makes sense for the bot to reset them at some point. — Andy W. ( talk · ctb) 00:16, 18 September 2016 (UTC)
Note that this bot is blocked to access webpages from the Minor Planet Center. E.g. this page.. see revert. I have already posted this fact before on this very page here. For the last few weeks, I have spent hours reverting these unhelpful bots edits and flagged the corresponding post on the talk-page with FAILED. This has to stop. Rfassbind – talk 20:46, 15 September 2016 (UTC)
is being forwarded to:
it is detrimental to label it as dead link and add an unhelpful, completely outdated archive link. The HTTP Status Code for the erroneously considered dead URI is a 302 found. Rfassbind – talk 00:30, 18 September 2016 (UTC)
I assume, youbot doesn't go trough all websites, right, but only on those, which are in your database? If so (and this post is useful), you can add *.london2012.com to list. If this wasn't useful post of mine (and bot archives really everything it can), say so :) --
Edgars2007 (
talk/
contribs)
19:10, 24 September 2016 (UTC)
In the edit your bot made ( diff link) three Alexa.com links got tagged as dead but where working when I checked. I should also let you know (if you don't already) that the Wayback Machine has had intermittent failures lately. Your bot should avoid substituting "https" for "http" within URLs or adding the protocol number ":80" (the Wayback Machine has done this) as another bot GreenC bot ( talk · contribs) cleans up such links as it did with the follow up edit. Regards. – Allen4 names ( contributions) 11:42, 25 September 2016 (UTC)
When IABot isn't provided the accessdate, what happens? Also does the IABot add archive urls to urls that aren't necessarily dead? -- Tim 1357 talk| poke 05:57, 25 September 2016 (UTC)
I was unhappy with this edit by the InternetArchiveBot.
I have reverted the edit, and tagged the article with a {{bots|deny=InternetArchiveBot}} template Hawkeye7 ( talk) 22:18, 26 September 2016 (UTC)
{{
cbignore}}
.—
cyberpower
Chat:Offline
22:32, 26 September 2016 (UTC)
{{
cbignore}}
documentation was unclear as to whether it would work in this case.Hawkeye7 ( talk) 23:03, 26 September 2016 (UTC)
{{
cbignore}}
works on anything the bot touches. It's parsing engine is quite intelligent. It's designed to mimic human parsing as best as possible.|dead-url=bot: unknown
is not intended for use by regular editors, it is only vaguely documented. It functions the same way as |dead-url=unfit
, used previously by this bot, except that it is semantically different and has its own maintenance category:
Category:CS1 maint: BOT: original-url status unknown. |dead-url=unfit
(or |dead-url=usurped
) is intended to be used by editors who find an original url that may once have supported an article's text but now has been taken over for the purposes of advertising, porn, malicious scripts, or whatever. When these kinds of urls are found, |dead-url=unfit
causes
Module:Citation/CS1 to hide that url while at the same time retaining it for historical purposes. This bot was using |dead-url=unfit
which has contaminated
Category:CS1 maint: Unfit url so now we don't know which of the 9k-ish original urls in that category are unfit and which the bot knows to be live but doesn't know if they support their respective articles.|df=
is documented for all cs1|2 templates that use {{
csdoc}}
. See
Template:Cite web#Date.There's a problem with this edit by the bot: For the access date, it inserted the date the archive was captured, not the (presumably) current date when the bot accessed the page. This date precedes the founding of Wikipedia, which correctly caused it to be flagged as an error. -- Floatjon ( talk) 15:58, 29 September 2016 (UTC)
I can't believe no one hasn't already asked you this but fer chrissakes adjust the bot so it stops spamming talk pages. The edit summary in the pages' histories is more than enough notice that the external links have been modified. Such notification is in fact precisely what they are there for. The talk page is for actual talk about the improvement of the article which "double checking bot edits" doesn't actually fall under. The only pages where anyone ever follows the bot's instructions, verifies the link, and makes a mark are well-curated pages where they would have checked the new links anyway.
At minimum—if there is some policy that requires such notifications—make them as minimal as possible, as a single text line without images: I have just added archive links to one external link on X. Please take a moment to review my edit. If necessary, add after the link to keep me from modifying it. Alternatively, you can add to keep me off the page altogether.
. Right now, it's an eyesore. Nothing else being written is important, creating
a category with 180,000+ entries neither you nor anyone else is whittling down is not helping anyone, and it is a needless piece of makework to require other people delete them by hand or craft other bots to follow around cleaning up after you. —
LlywelynII
02:42, 27 September 2016 (UTC)
|checked=failed
, thinking that it might be useful feedback. I think myself a little unusual in going to such trouble. On even my most active articles, I have little competition to complete this task. I neglect to change archived messages on pages I'm new to, even if I check the link, since changing archived material is generally discouraged. The failed links that go undetected should be either relisted or reworked with time, especially since the bot seems to work on automatic pilot in checking for failed links.
Dhtwiki (
talk)
23:38, 29 September 2016 (UTC)
on the 2 Wayback sources you added here. a couple of hours ago (I think). Could you please take a look? Thanks, -- Hordaland ( talk) 04:32, 1 October 2016 (UTC)
Thinking about {{
official website}}
(aka {{
official}}
), see this
for example. This is a complex template as it gets link data from Wikidata and compares it with the source link, and if they don't match it considers it an error and adds to a tracking category. In this example that's what happened. I can see three possible solutions:
{{
wayback}}
to the end eg: *{{Official|example.com}} {{webarchive |url=https://web.archive.org/web/*/example.com |date=* }}
*{{webarchive |url=https://web.archive.org/web/*/example.com |date=* }}. In this case it would also have to transfer any
|name=
argument to wayback's |title=
.Each method has pros and cons. Official is a widely used template with 162192 transclusion, more than double the number of wayback templates. Probably the least disruptive would be #2 because it doesn't touch the original official, though it presents problems if the archive isn't templated like wayback or webcite. #3 would be the best performance since it reduces number of templates. #1 would be smooth but increases complexity of that template code and unclear there is consensus for it. And it will be up to you what is better solution for the bot. -- Green C 13:32, 30 September 2016 (UTC)
This nested a dead links template inside another template creating an error on render. -- Green C 13:08, 30 September 2016 (UTC)
{{
official}}
template, because it is neither an archive template or a cite template, nor does it share any cite template functionality. So it just sees the link inside it.—
cyberpower
Chat:Online
12:21, 1 October 2016 (UTC)
Hi! In this edit, InternetArchiveBot added "deadurl=no" and an empty "df" parameter. First of all, the original URL was dead when I tested it ("server not found"), so the parameter should have been "deadurl=yes". Second, why the empty date-format parameter? Is that supposed to be filled in later? Thanks. — Gorthian ( talk) 00:52, 2 October 2016 (UTC)
I just noticed on WP:RFPP that if a request is not [ signed Cyberbot I will make an edit saying it can't be parsed. Scroll down and there is a error message caused by the lack of signature in the original request. This causes the bot to return and tell itself that the previous bot message can't be [ parsed. Don't know if you have some way to suppress the error message in the first bot comment or stop the second. Cheers. CambridgeBayWeather, Uqaqtuq (talk), Sunasuttuq 09:25, 2 October 2016 (UTC)