[07:29] * Nemo_bis still downloading shoutwiki.com at 5 KB/s max :-( [12:07] OMG, http://icehockey.shoutwiki.com has 80 000 pages [12:31] sounds like spam [16:10] 80,000 just for ice hockey? that is madness. [16:12] indeed [16:13] I had to split the task, now I have an instance only for that wiki [20:39] english wikitravel finished, but no images [20:47] by the way, most images are on this shared project http://wikitravel.org/shared/Main_Page [20:48] ah, nice [20:48] why can't they just use Commons? [20:48] Do they have many unfree images? [20:49] maybe [20:53] is it allowable within Commons' policy to upload images that are *intended* for use outside of Wikimedia Foundation projects? [20:54] i dont think so [20:54] but if they are on Commons scope, maybe [20:57] We are glad to announce the 100th wiki dump of WikiTeam: English WikiTravel: http://code.google.com/p/wikiteam/downloads/detail?name=wikitravelorg_wiki_en-20110605-current.xml.7z [21:01] i am running a script to automatically generate a torrent for the upcoming enwiki-20110722-pages-articles.xml.bz2 [21:03] 1) it checks the RSS on dumps.wikimedia.org every 60 seconds, 2) as soon as the new file is published, 3) it feeds the http link to burnbit.com, which generates a torrent [21:04] i'm using Yahoo Pipes to generate a search-based RSS feed from burnbit.com, which i have plugged into my torrent client for automatic downloads [21:05] why don't you use the 7z one? [21:05] pages-articles is only in bz2 [21:05] ah, ok, page-articles [21:05] pages-meta-history is in bz2 or 7z [21:05] average size? [21:05] here's a link to the Pipe, if you're interested http://pipes.yahoo.com/pipes/pipe.run?_id=f55afdb4a562a2a7ec5096624ac9a4e6&_render=rss&q=enwiki+pages-articles [21:06] i assume the next enwiki pages-articles will be slightly larger than the last, which was 6.92 GB [21:06] interesting, can you do that with hhttp://code.google.com/p/wikiteam/downloads/list ? [21:07] yes, i'll have it done by tomorrow [21:07] paste your links here http://archiveteam.org/index.php?title=Wikiteam [21:07] rgr [21:08] i've used Yahoo Pipes to take existing Podcast feeds (with http links), then automatically convert everything into a .torrent link using the Burnbit service [21:09] LobStoR, yes, it's certainly allowed. Commons' scope is pretty broad [21:13] downloading the last page-articles torrent [21:25] * Nemo_bis is using up all bandwidth for JSTOR torrent [21:27] torrent feeds would be good for distributing wikiteam/archiveteam data automatically [21:27] i'll definitely work on that [21:37] i think internet archive could add a link to all their items to torrent, using ther server as webseed [21:41] hmmm, the pipe i made a few days ago is now broken, due to Burnbit adding a captcha [21:41] they must have thought it was some sort of bot [21:41] which, i guess, it sort of is [22:12] ok, emijrp... i still have some tweaks to do, but the wikiteam torrent feed is ready [22:12] i've tested successfully it in utorrent on windows (which downloads from the "enclosure" URL) http://pipes.yahoo.com/pipes/pipe.run?_id=2f15d291dd975362a2dfdb1da3c3ba4d&_render=rss&project=wikiteam [22:15] Google Code uses atom feeds, so i still need to adapt some of the elements to work properly on RSS (such as the dates), but it works without them, for now [22:27] "pipes.run" keeps triggering the spam blacklist on the wiki for ".ru" :-( [22:30] are all .ru domains blacklisted? where? [22:35] archiveteam.org [22:35] not Wikimedia [22:35] http://archiveteam.org/index.php?title=Wikiteam#BitTorrent_downloads [22:35] ah so not my fault :-p [22:36] haha, yes, one way of putting it [22:36] I can't manage to download those torrents [22:36] Are all torrent clients able to download from webseed? [22:36] eg Transmission [22:36] there's two different webseed standards [22:36] and not every client supports it [22:36] transmission doesn't support the webseed method on those torrent [22:37] give it a few minutes, and me and emijrp should be able to seed it to you ;-) [22:38] ok [22:39] it's a shame that there's two different webseeding standards, of which neither is really very well implemented http://en.wikipedia.org/wiki/BitTorrent_%28protocol%29#Web_seeding [22:42] perhaps we can coax the folks at Burnbit to support both webseed standards in their .torrent metafiles, so that they aren't alienating users... but for now, this is what we're stuck with [22:56] perhaps they're trying to make their standard more popular so that it becomes THE standard [22:57] One webseed standard requires a cgi script to run on the webseed server [23:01] Yahoo Pipes has very aggressive server-side caching, which makes sense [23:02] but makes it difficult for me to confirm if i finally got the dates working properly, since i have to wait 15-30 minutes for the feed to reflect any changes to the Pipe [23:02] I wonder why they shut down Geocities but keep such a strange service [23:08] yeah, pipes probably really doesn't even bring them any revenue, since the whole point is to generate a feed... resulting in no pageviews on yahoo.com [23:08] unless they start inserting text ads into your feeds :-P [23:18] small filesizes (< 1 MB?) are blocked on Burnbit, which is unfortunately many of the Wikiteam downloads