#wikiteam 2014-01-25,Sat

↑back Search

Time Nickname Message
02:13 🔗 SketchCow I think you know
04:01 🔗 Danneh__ whoo, backing up wikis is fun
07:52 🔗 Nemo_bis Danneh__: I'm glad you think so! :)
07:52 🔗 Nemo_bis did you already upload some to archive.org?
07:58 🔗 Danneh__ Nemo_bis: Yeah, just a few game ones. Downloading Combine Overwiki now, about 12720 pages in (been running for two-three days so far!)
07:59 🔗 Danneh__ Around Should be done today or tomorrow, depending on how long it takes to grab the images
08:02 🔗 Nemo_bis Danneh__: nice, and you know about uploader.py right?
08:03 🔗 Nemo_bis if you use that one, you easily add all the metadata and then, for instance, wikiapiary.com will automatically create a page on that wiki (which they don't have) and link the dump
08:03 🔗 SketchCow aa/iwin 14
08:04 🔗 * Nemo_bis just *knew* someone was going to use the metadata in the end
08:05 🔗 Danneh__ haven't messed around with much in the way of metadata yet, to be honest
08:05 🔗 Nemo_bis SketchCow: I use this finnish trick http://niklas.laxstrom.name/page/eng/irssi
08:05 🔗 Nemo_bis Danneh__: one more reason to let the script handle it :P
08:05 🔗 Danneh__ I know if I throw a list of URLs in a text file and point uploader.py to it, I can get them on archive.org, haven't yet looked into it more in depth yet :P
08:05 🔗 Nemo_bis let me know if you have problems with it, it's quite hacky still
08:06 🔗 Nemo_bis I just added some docs for you ;) http://code.google.com/p/wikiteam/wiki/NewTutorial#Automatic_publishing long overdue
08:06 🔗 Danneh__ ah, thank you very much!
08:07 🔗 Danneh__ I will need to add that --help option and submit a patch, but been super busy lately
08:10 🔗 Nemo_bis sure
08:11 🔗 SketchCow NICE INNOCULATION SCAR, PERFECT ROBNOT
08:14 🔗 Danneh__ just wondering, if they're using Semantic MediaWiki, know if that's currently backed up using dumpgenerator.py ?
08:22 🔗 Nemo_bis Danneh__: SMW help pages say that all the properties and stuff are stored on the wiki
08:23 🔗 Nemo_bis what I'm not sure about is what happens with features using a different contenthadler
08:23 🔗 Nemo_bis contenthandler
08:24 🔗 Nemo_bis hmm seems straightforward enough, no real problems https://meta.wikimedia.org/wiki/Special:Export/Schema:MediaWikiVagrantUsage
08:25 🔗 Nemo_bis though that adds to the complexity of the installation configs... we save special:version but that's hardly enough
08:26 🔗 Danneh__ fair enough, was looking into a wiki that uses SMW, wanted to make sure before I back it up
08:26 🔗 Danneh__ well, before I try at least :P
08:28 🔗 Nemo_bis :)
08:29 🔗 Danneh__ just using home server, 'bout 10 seconds delay, hasn't given me any problems yet
08:29 🔗 Danneh__ :)
09:30 🔗 Nemo_bis ah yes that's slow :)
09:59 🔗 Danneh__ aha, slow but safe, don't wanna get myself banned on here!
10:00 🔗 Nemo_bis the trick is to be fast enough that they don't notice in time :P
10:00 🔗 Nemo_bis does someone want to extract all the domains of their wikis from the page in http://www.shoutwiki.com/wiki/Category:Flat_list_of_all_wikis ? we need to update https://code.google.com/p/wikiteam/source/browse/trunk/listsofwikis/shoutwiki.com
10:04 🔗 Danneh__ aha, fair enough
16:36 🔗 Nemo_bis https://archive.org/details/wikia_dump_20140125
21:08 🔗 Nemo_bis I'm so tempted to just start download of a few thousands wikis on my shared server...

irclogger-viewer