[01:30] Back [01:35] God nap [01:35] THe nap of the gods [04:30] SketchCow: teach me how to nap like a god [04:30] I barely get enough sleep as it is [05:25] BlueMax: 90 minute intervals [06:54] "archive me.com directly to archive.org... ....it's better to ask about this way on our IRC channel. " [06:57] why is it better?^_^ im gonna try now to fill a gbit... [07:06] ------- [07:06] You can't press any keys yet, it's slow, no sound either. But JSMESS has had a breakthrough: [07:07] ------- [07:07] http://interbutt.com/temp/jsmess_cosmofighter.html [07:20] Interbutt, brilliant website name [07:20] Aaaaand the site crashed Firefox 10. Brilliant. [07:21] Wait, it unfroze. [07:53] jonas__, it's better because otherwise we have no way to easily pack and upload you stuff to archive.org [07:53] try the standard script first, we'll see how much you download ;) [08:23] holy crap [08:23] also, no wonder my computer is warm and slow now [14:23] Yeah [14:23] Work to go! [14:27] Going to try and activate ReCaptcha again on archiveteam.org. [14:31] OK, according to the ReCaptcha page, the ReCaptcha plugin is abandoned. [14:38] That sucks [14:38] I bet SketchCow just had a FFFFFFFUUUUUUUUU moment. [14:38] No, I really, didn't. [14:39] Wow, you're back. Is school out? [14:39] SketchCow: you talking to me or ersi? [14:43] BlueMax: You [14:44] I'm always here ;) [14:44] lol [14:44] It's actually stupidly late here [14:44] And I've been around for weeks [14:44] Just haven't said anything [14:44] Because around you people I feel like the dumb guy who can't turn his PC on [14:47] http://xeducation.info/2012/03/young-fresh-czech-teens-hard-fucked-in-all-holes-and-position-full-movie-jeste-jsem-panna/ [14:58] Anyway I must be off, toodle-oo and I'll try to talk here more often so SketchCow doesn't think that I'm a skeleton in front of a computer, bye now! [15:04] I was really, really hoping the fucked in all holes thing was a legit archive team project [15:05] "Who wants in on the young fresh czech teens hard fucked in all hol" [15:08] I wonder who the lucky researcher will be that writes "30 Years of Internet Porn" [15:08] Lucky, heh [15:09] Maybe not so lucky if he's watching 30 years' worth of porn [15:19] so mess.org is down, let's play guess the webhost response: a) oh our bad the storage array should be rebuilt in an hour or two, b) oops please sit tight while we go fetch the backup media, c) backups what backups [15:27] DFJustin: it's still down? :< [15:28] Is wikiteam downloading that wiki [15:28] (hint it's c) [15:29] it's a dokuwiki and not a mediawiki so you'd need a different tool [15:29] So.. mess.org made.. a mess? [15:29] apparently the webhost did something [15:29] hopefully it wasn't dreamhost… heh [15:30] don't think so [15:39] I want to people to be able to do projects involving project gutenberg text easier. [15:41] I see lots of art and demo projects on hacker news, I can't help but think if people could get all of the PG texts without having to use rsync, and then trying to parse the files (file names are not book titles) [15:41] Getting the archive, checking it into git, and trying to get github to host it seems like a reasonable first step. [15:42] Does this sound like something archive team is into? [15:45] sure [15:45] um I dunno if guthub will want 600gb of text or whatever it is [15:46] yea, I don't think they have a plan that large [15:46] you don't need to parse the files for book titles though they have a database http://www.gutenberg.org/wiki/Gutenberg:Feeds [15:46] but you can use git annex [15:50] Dreamhost JUST shut down urlte.am [15:50] For having "malicious data" [15:50] I guarantee you, it has had malicious data for a while. [15:52] Odd. I didn't think that PG wanted people accessing their system with scripts/robots [15:52] but this is basically an API [15:52] http://www.gutenberg.org/wiki/Gutenberg:Information_About_Robot_Access_to_our_Pages [15:53] I've seen that page. [15:54] but missed the important link at the bottom [15:54] Does anyone know how PG takes the output of distributed proofreaders and turn it into epub, txt and html? [15:54] They presumably have some sort of build script [15:54] yeah, I think it is public somewhere or other [15:55] http://www.pgdp.net/wiki/PPTools [15:56] the txt and html files are what people produce, other stuff like epub is autogenerated by gutenberg.org [15:57] DFJustin: Thank you, I could not find that last night. [16:04] https://plus.google.com/u/1/101174951617223562800/posts/2rk37bdDyep [16:04] "Piles of"...? [16:05] Hard Drives [16:05] As you see, they had fill in the blanks. [16:05] The top one was "My favorite App is..." [16:21] http://www.archiveteam.org/index.php?title=Special:RecentChanges [16:21] If you go there, do you see "what year is it again" [16:22] yes [16:25] should have something on there about fortunecity [16:49] Hi. Who is "highscore" on the FortuneCity tracker? [17:34] SketchCow: [17:34] [1:31pm] LordNlptp: so the redump wiki, ftp and ALL the files on it were destroyed due to disk failure or software error [17:37] hang on you haven't taken kleptomania into account [17:37] this is true [17:38] I have the ftp contents at least [17:47] Wow. [17:51] Want to mirror them? [17:51] I can put them in a dark item on archive.org [17:51] Or not dark, not in the search engine. [17:52] SketchCow: is there any way for you to dig out the latest wiki from wayback? [17:52] and/or bugzilla if it's there [17:52] yeah probably worth a go, I'll need to do some basic organization first [17:52] in the meantime it's on 3 drives in 2 locations [17:54] http://wayback.archive.org/web/*/http://www.mess.org [17:54] I show the last backup as July 23, 2011. [17:54] Which is a shame, but is something. [17:55] yes, but doesn't wayback have a 6+ month delay? [17:55] I can ask. [17:55] also http://wayback.archive.org/web/*/mess.redump.net [17:56] err, the latest mess.org one is the old wiki [17:56] :/ [17:56] or old site [17:56] the mess.redump.net wiki is what we're after. [17:56] ? I looked earlier and it was the wiki [17:56] http://web.archive.org/web/20110724031059/http://mess.redump.net/ [17:57] wait, maybe I clicked on a slightly old one [17:58] I don't think there have been a ton of edits since july, the main thing is missing pages [18:00] (by way of explanation, both mess.redump.net and mess.org were pointing to the same site) [18:07] Yeah. [18:08] archive.org doesn't have an easy way to see crawls "in the pipeline", so to speak, I've been told. [18:09] :< [18:20] lord_nigh's busy pulling stuff off google cache, wasn't there a site out there to automate getting you all the available caches of something [18:22] ah yes http://warrick.cs.odu.edu/ [19:29] balrog: Redump got hosed? Yikes [19:29] shaqfu: yup, hosed [19:30] balrog: THank god for UG, then? :P [19:30] does UG have the pages? [19:30] the wiki is what I'm worried about [19:30] Ah, nope [19:30] Just the game files [19:32] There's something bitterly ironic about a community dedicated to saving old software not keeping backups of itself... [19:32] I know… [19:34] redump.net != redump.org [19:34] DFJustin: yes, redump.net [19:34] not redump.org [19:34] and well MESS is a community dedicated to saving old hardware/stuff and even software [19:40] Giant Bomb has sold to Gamespot [19:43] Wasn't it formed by disgrunted ex-Gamespot writers? [19:48] Yes [19:50] :trollface: Problem? [20:19] SketchCow: are you coming to northern germany? :) [20:21] willwill: Are you around? [20:21] yes? [20:21] Are your mobileme downloaders allright? [20:22] You seem to have claimed 120000 usernames. [20:22] ah yes, but I haven't sshed to the download machine for few days [20:22] well, I ran seesaw and my disk is full [20:22] Ah. [20:22] can you clear all the username I claimed? [20:22] ahh, makes sense :) [20:22] Yes. [20:23] willwill: Should I also reset the usernames that you have marked 'done'? Or are those safe? [20:23] Username marked done should be already uploaded as I use seesaw.sh [20:23] If he ran seesaw, those should be fine [20:23] Good. Thanks. [20:23] nice nice [20:24] alard: is this why it shows 122953 to do? [20:24] :p [20:24] Probably. 'todo' does not include 'out'. [20:35] I home I will [20:37] alard: btw, detecting disk nearly full and stopping would be a very nice feature for your framework to have.. [20:46] closure: Yeah, perhaps. (But it is on github... :) [20:46] heh yes [20:47] maybe 3x avg item size [20:54] http://img.pr0gramm.com/2012/03/ww6y1.jpg [21:04] alard: well, my attempts so far are crappy shell loops that run df . and touch STOP [21:14] ersi: hahahahaha