[19:02] I'm trying to use dumpgenerator.py to archive http://www.frathwiki.com/, but I'm getting "Error in api.php, please, provide a correct path to api.php". The command I'm running: python ../dumpgenerator.py --api=http://www.frathwiki.com/api.php --xml --images Any thoughts? [19:26] kyan: try adding also the --index= [19:27] sometimes it's stupid enough to be the reason, though it shouldn't because it's in the same path http://www.frathwiki.com/index.php [19:32] no idea why it would fail https://code.google.com/p/wikiteam/source/browse/trunk/dumpgenerator.py#901 [19:33] UA blocking [19:33] Access denied | www.frathwiki.com used CloudFlare to restrict access [19:33] oops [19:34]

The owner of this website (www.frathwiki.com) has banned your access based on your browser's signature (bad82647cc6077f-ua48).

[19:36] it's more than just UA [19:36] Nemo_bis: ^ [19:38] https://support.cloudflare.com/hc/en-us/articles/200170086-What-does-the-Browser-Integrity-Check-do- [19:38] hmm [19:39] wget and curl work [19:40] yes, that's what confused me :) [19:41] are they blocking urllib completely? [19:41] they're finding some way to block it [19:42] I changed the first two lines of checkAPI to the following and it works: [19:42] f = urllib2.urlopen(req) [19:42] req = urllib2.Request(url=api, headers={'User-Agent': getUserAgent()}) [19:43] getPageTitlesScraper may still be broken [19:43] they're probably blocking the urllib UA [19:43] yeah [19:43] strange… [19:43] I was doing the same change at the same time [19:43] and the script isn't consistent and is using urllib in a few places and urllib2 everywhere else [19:44] indeed [19:49] ok, fixed that one (thanks balrog), only two urllib.urlopen left :) [19:49] probably should fix those as well or those gets will fail [19:51] yeah but they're there only for the older wikis, this one shouldn't be affected [20:02] The new version of the script is working now :) [20:02] thanks! [20:14] kyan: thank you for reporting, it was a rather embarrassing overlook :)