[02:11] Now, let's destroy some backlog [02:12] How'd jjonas find the new usernames? [02:12] alard: mothony 9361MB [02:12] oof [02:13] iam just doing excessive crawling for the last few % [02:21] spam: http://archiveteam.org/index.php?title=Information_About_PPI_-_Payment_Protection_Insurance_201206 [02:22] You don't say [02:23] Debianer: That's impressive, given that signups have been broken [02:23] No, they've been fixed for 3 days [02:24] next, I'm adding a "confirm new user" step for admins [02:24] That didn't take long to get spam [02:24] Which I and others will take on [02:24] How do you think spambots work? [02:25] I know, I've seen access logs. [02:25] Fetch a page (e.g index.php?title=Car_Insurance_get_cool_62) [02:26] GET the associated edit page and POST some spam text with it [02:26] (Optionally) Create an account and upload some photo, or make an article [02:32] http://www.mediawiki.org/wiki/Manual:$wgEnableDnsBlacklist [02:33] afaict, that blocks new user creation. [02:33] i could be wrong, though [04:51] is the memac tracker down? i keep getting an error when marking a user completed. [04:56] the dashboard works [04:57] not sure if the trackers the scripts talk to are responding [05:00] oh it started working now [05:00] got about 10 errors until it went through [05:00] might have been a hiccup somewhere between you and it [05:00] hmm [05:01] I wonder if the seesaw script re-randomizes on each notification retry, or if it keeps retrying the same node [05:02] I see several 0MB entries on the recent completions list [05:14] lemonkey: yes heroku was down [05:14] we're back in action now though [05:14] lemonkey: actually all of us-east ec2 was down [05:15] holy shit [05:16] they had to change the oil on the second backup genset [05:16] all of the entire AWS zone was down. that's pretty big [05:16] Coderjoe: EBS failures on one AZ + api latency for the whole region [05:17] oh that's just awesome [05:17] Coderjoe: so fun! [05:17] can't wait for the tech news/blog coverage [05:17] https://status.heroku.com/incidents/375 [05:17] everyone was down [05:18] Pinterest, Quora, AirBnB, DotCloud, Parse, HipChat, Svtle, etc [05:18] lol [05:18] ~cloud~ [05:18] all the hipster companies [05:18] 1/3rd of netflix [05:18] netflix instant [05:18] that is one outage the crazy monkey doesn't prepare you for [05:19] well, so much for 'redundant' [05:20] Coderjoe: chaos monkey *is* ec2 ;) [05:22] :P [05:27] http://status.aws.amazon.com/ [05:27] mm [05:27] all those little yellow triangles [05:28] Coderjoe: they were all green until 20 minutes ago [05:28] took them 45 minutes to turn them yellow [05:36] "We can confirm a portion of a single Availability Zone in the US-EAST-1 Region lost power." [05:39] the clown stepped on a banana peel [05:41] the maintenance crew had to plug in their waxer [05:48] someone tripped over the linksys router chord [05:48] I know a guy that worked at a company with a good-sized datacenter in Sacramento, CA. One day, a contractor was doing work, bent over, and managed to hit the estop button with his ass, taking down most, if not all, of the datacenter [05:49] * Aranje laughs [05:49] I was once a foot away from hitting the estop button on one of microsoft's maps clusters [05:50] http://www.southparkstudios.com/clips/166192/internet-reset [06:16] SketchCow: is that the 24 1.4L? [06:27] No [06:27] I know you're all in love with your news [06:30] SketchCow: indeed :) [06:31] hopefully won't need any more glass… ever [07:02] * SmileyG doesn't tell Coderjoe about the power strip they have on the floor in their small server room, which someone trod on ¬_¬ [07:09] mp3 [07:17] dudes dudes dudes got an issue.... [07:20] how exactly does the --no-clobber switch work in wget, i was running get on a remote machine when my local machine crashed closing the ssh session when i rebooted and logged back in the wget session was no longer running in the back ground so i started it again.... am i doing right? will using --no-clobber work out? [07:21] --continue --no-clobber are happy bedfellows when resuming crawls [07:22] you need to run things inside of `screen', though. [07:23] have been, i like seeing what going on, should i stop and added --continue or leave it going with just --no-clobber ? [07:23] (type fail) [07:23] --continue will make it resume faster [07:23] what are you mirroring? [07:24] http://planetquake.gamespy.com/ [07:25] \o/ [07:25] lol hey Schbirid [07:25] had a little freak out just now Schbirid read up ^-^ [07:26] seems to of resumed ok though :D [07:27] cant read up, i just joined :) [07:27] weird that it is so slow for you, i was done in a couple of hours. but maybe you actually get the pages i was missing [07:28] summary, local machine crashed, took ssh/wget session with it, just resumed [07:28] how big was it when you 'completed' it ? [07:28] you shall use screen! [07:30] how bigggggg [07:31] 970M [07:31] dude.... im at 9.1GB !!! [07:32] :O [07:32] what are your wget options? [07:32] grabbing all the linked zips and other random content/mods linked too, alot are hosted elsewhere and in sub domains [07:32] ah [07:32] yeah, i planned that for later [07:32] nice [07:33] may aswell get it done, where am i uploading when done? [07:33] before uploading, did you check for the file i was missing as example? [07:33] nor does it download the page "http://planetquake.gamespy.com/fms/Image.php?image=http://pnmedia.gamespy.com/planetquake.gamespy.com/images/oldsite/clusterimages/casspq1.jpg" [07:33] the page "http://planetquake.gamespy.com/View.php?view=LOTW.Detail&id=150" is downloaded but not the image http://pnmedia.gamespy.com/planetquake.gamespy.com/images/oldsite/clusterimages/casspq1s.jpg [07:33] yup, the files you told me about yester are there [07:34] wicked [07:34] what options do you use? [07:38] (see pm) [07:39] also this - http://i.imgur.com/He19v.png [08:38] Herp, musicksmonument on MobileMe causes OOM on warrior [08:44] SmileyG: #memac for MobileMe. Also, that happens for a lot of users - you'll need more memory to complete that user. If you don't have that (for example, in the Warrior) just leave it be and another downloader will be allocated that user [08:46] We're stuck with a set of users that have been carefully selected to crash almost any downloader. [08:48] lol, seriously [08:53] Yupp, it's the 'end-game' [08:55] Hm~ cloning CD's, would 'dd' be alright? Or should someone engross in some magic dances with extra tools and such? [08:56] cdparanoia [08:56] * ersi searches for it [08:56] ooh, I see [08:56] 's good shit [08:57] This seems more focused on Audio though [08:57] I'm all Data, maan [08:58] dataman [09:03] I'll just dd this bitch, it's better than nothing [09:03] dd the shit out of it [09:10] https://en.wikipedia.org/wiki/Razor_1911 got speedy deleted because of "A7: Article about a group or club, which does not indicate the importance or significance of the subject", but reading that rule, it does not apply: "The criterion does not apply to any article that makes any credible claim of significance or importance even if the claim is not supported by a reliable source or does not qualify on Wikipedia's notability guidelines." [09:10] the page in google cache is very short, but the image has a lot of contents! my guess is a lot of valid content was removed before deletion [09:11] what are the proper channels to bring it back? [09:12] Enjoy going into a jerkoff war at Wikipedia [09:14] A lot of things gets deleted from Wikipedia, if it's content that's not speed deleted right after it was created - it'll be in the dumpfiles [09:23] lrkj: setup wikipedia2.com; go to town restoring all the backups [09:24] in that regard, wikipedia is a bug clusterfuck [09:26] And in that regard, we are going offtopic ¬_¬ [09:27] already were, really :-\ [09:35] archiving CDs are quite on topic [09:41] but all the wiki stuff? [09:41] #wikiteam is an official subcommittee [09:42] ersi: ddrescue might do better than dd [09:43] which ddrescue? the gnu one or the other one? [09:43] Dang, there's another one? [09:44] yes, there are two [09:44] Evidently. [09:44] I've used the non-gnu one more [09:44] Well, I guess either is an improvement over plain dd. [09:44] I used brasero :$ [09:45] bleh. i need to try and get some sleep. I suspect I'll be getting a call in 3-4 hours :( [09:47] Is there a good CD imager that can also save copy protection stuff like subchannel data and whatever? (As you can see, I'm an expert on the topic.) [09:57] Deewiant: cdrdao can read subchannels [09:58] Thanks, I'll look into it. [10:01] brasero front-end uses cdrdao [10:06] ersi: if you find something "always good" for cd ripping, please tell me [10:07] of course. I'm not really actively looking though [11:06] audio cd? [11:06] data, audio, data+audio, *+copy protection [11:06] btw theres dd_rescue, and ddrescue :D [11:07] Schbirid: I've not done it before but I'll keep an eye out. [11:07] and dd into an .img would be my bet. [11:07] then you mount it mount -o loop ./cd.img ./mount/point/ [11:08] I don't know if dd cares about the physical disk tho, and if the copy protection does either. [11:08] I presume it does. but I've not tested. [12:45] 15 minutes until today's ovh server giveaway [12:54] 7 minutes [12:56] 4 [12:59] nao? [13:00] success [13:00] i think i got 5 codes in total now, only managed to properly register one :B [13:00] Um, that was quick. [13:00] It took 2 seconds. [13:01] er, 4 [13:02] hm, not getting the DM [13:15] Schbirid: I wouldn't mind a code! [13:18] GLaDOS: they are useless or do you know where one can enter them if the form is gone? [13:19] i better wireshark next time [13:19] Nah, no idea how to enter them. [13:19] :\ [13:22] Whew [13:22] OK, let's see what we can do today. [13:35] SketchCow - I've heard there's audio of the JCDL talk, you should archive that and then give me the link so I can back it up in my ears [13:54] Hello, I just got an email that wakoopa is shutting down and thought you guys might be interested. More details: http://blog.wakoopa.com/post/24878499948 [14:10] It's not a good recording. [14:13] I'm not picky [14:14] besides, when did crap quality mean you shouldn't share? :D [14:17] I'm not sure that previous comment really put across what I meant to say [14:18] regardless, I still think you should share it [14:34] Uploading now; [14:34] SketchCow: if I start ripping my data CD collection (and I have tons of various CDs, including old shareware, software from now-defunct companies, and more), how would I go about uploading it? [14:34] I have to have several hundred CDs [14:35] SketchCow - sir, you are awesome... not as awesome as your hats, but awesome nonetheless [14:35] An easy way to to scan them and the stuff, and I provide an FTP and can do all the ingestion. [14:35] 25% done on upload of speech [14:35] scan the disk itself, and any papers in the cd case, right? [14:37] sketchcow eats shareware cds for breakfast [14:38] DrainLbry - most people would eat the cereal, but the free gift is probably just as nutritious [14:41] any particular reason there's not a Hornet Archive copy on IA? I was a bit surprised. Looks like Wayback crawled scene.org, but the file links point offsite so it doesnt have them. [14:41] If not any reason, I'll put it on my to-do soon list [14:42] hornet? [14:42] Demoscene archive [14:42] ah neat [14:42] Which apparently disappeared off the face of the earth at one point years back, and from what I read looks like some may have been lost. Surprisingly small number of mirrors of it i've dug up. [14:43] http://archive.org/details/JasonScottKeynoteJCDL2012 [14:43] Also has been the victim of Wikipedia's revisionist "non-notable" article deletion policy. [14:43] * ersi rolls eyes at Wikilolpedia [14:43] Do you mean the Hornet CDs? I have those up, I thought [14:43] sketchcow: yeah, i didnt dig them up in a search [14:44] was surprised [14:45] SketchCow - ta muchly [14:47] there seems to be some rendered demoscene video, but i cant find the hornet collection as a whole, or the original executable democode [14:48] I put a few democene video dvds up on archive.org [14:51] sketchcow: i couldn't help but think "the clown went down" when I saw the AWS outage article this morning. it's stuck in my head now! [14:58] also, there's some hillariousness to be drawn for the tracker site used by AT warrior/seesaw, which runs on AWS infrastructure, going down due to cloud outage [15:11] sketchcow: found two hornet discs on cd.textfiles.com , seems to be some differences if you were to mash these two disks up with what's on scene.org - the 2 CDs, and their collection have differences. [15:11] looks like I found myself a project. The "definitive hornet archive" [15:43] Time to mirror Apple's Ping - http://www.itproportal.com/2012/06/14/apple-shut-down-music-social-network-ping/ [15:43] " [15:43] "Ping, Apple's laughable attempt at a social network, will be shut down with the next major release of iTunes, according to a report. [15:43] Read more: http://www.itproportal.com/2012/06/14/apple-shut-down-music-social-network-ping/#ixzz1xsRo5VzP [15:44] is it me or is more and more stuff closing down [15:47] SketchCow - that talk sounds alright to me, no worse than any other hacker con video out there [15:48] which probably doesn't sound much like high praise, but it's perfectly listenable [15:49] It's not high praise [15:49] Most are shit [15:49] sketchcow: i'm also listening, sounds good [15:49] indeed, but they serve a purpose [15:49] and it's not to demo audio gear [15:50] Poor attitude. [15:50] That's like saying a rusted car is "not to demonstrate chrome" [15:50] firing up iTunes/Ping to see if I can figure out some URL formats, etc [15:50] Since it refuses to open in a web browser [15:50] well, I see your point, but at the same time, does that mean we should throw unique media out because it doesn't soud like it was recorded in a pristine, multi-million dollar studio? [15:50] I wanted to hook to the mixer but the facility they held this in had some insane per-item cost, i.e. for the microphones, mixers, time, etc. So to have me hook mine up would have been significant cost. [15:51] *sound [15:51] that sucks [15:51] Are those our two choices? [15:51] Deleted or pristine? [15:51] No. [15:51] You're framing ad absurdum [15:51] I was providing a counterpoint, but you've already proven there's a middle ground by uploading that file [15:52] I'm just saying that to some extent, unless it's horribly intolerable and near impossible to extract the content, I don't think quality is necessarily that important [15:52] helpful, sure, but not critical [15:53] I don't have a problem with "it's the best we can do" [15:53] and I've heard a hell of a lot worse [15:53] But as I found out, they hadn't given a thought to saving anything, so I happened to have it. [15:53] But let me say. [15:53] My #1 pet peeve? [15:53] besides underscor of course [15:53] Is "Oh, well, it could be worse" [15:53] I.e. "Well, let's let shit define our quality" [15:54] "Look, OK date, at least it wasn't rape with a broomstick" [15:54] "Car ride was OK, nobody was killed" [15:54] that's not what i'm saying, but there's something to be said for some tolerance to mediocrity [15:54] Sure it is. [15:54] You're pointing at shit and going "it's not THAT" [15:55] that's not necessarily bad, but what I'm saying is that it's above my threshold of "awful" [15:55] I can extract the content without giving myself a headache trying to concentrate it, and, for the sake of extracting that content, it's perfectly adequate [15:56] that, to me, is sufficient [15:56] not perfect, but sufficient [15:56] Jeez, you have world-class standards there. [15:56] WORLD. CLASS. [15:56] Move over Michelen [15:56] Meche;lng [15:56] Michenlin? [15:56] The tire guys [15:56] Michelin? can't recall off-hand [15:57] I'm just not that picky [15:57] I prefer quality, of course [15:57] but if it takes tolerating mediocre standards to get at something, well, that's ok too [15:57] there's an awful lot of shitty media out there with gold hidden in it [15:58] he's saying that he prefers the only copy of the water damaged book with no binding and ripped pages than no book at all [15:58] which you wouldn't see if you couldn't tolerate the abysmal VHS rip it came from [15:58] yeah, what Famicoman said [15:58] I'm not firing on all cylinders today, my metaphor and similie generator is offline [15:58] but Famicoman is exactly right [15:59] My interest in this discussion has ended. [15:59] well, then let Famicoman's correction of my terrible explanation of my stance be the final word [16:00] because he summed it up far more succinctly and accurately than I was able to [16:03] Ping uses user-agent to determine if you're allowed in , user agent that worked for me stolen from iTunes: iTunes/10.6.3 (Windows; Microsoft Windows 7 x64 Enterprise Edition Service Pack 1 (Build 7601)) AppleWebKit/534.57.2 , Example Ping Band Profile URL - https://c.itunes.apple.com/us/profile/id-10487 , example USer Profile: https://c.itunes.apple.com/us/profile/id181800443 - User ProfilesID value in URL seem sequential , returns an error if t [16:03] he Apple ID is not set up for ping [16:04] Ping's going to be a pain. [16:04] Does Ping have user data? [16:04] its got twitter style updates [16:04] and user profiles, so yes [16:04] Well, take a shot at it, good practice [16:04] Ping's shutting down, eh? Guess I'm not too surprised. [16:05] What's the download program that destroys limits on download speed again? [16:06] The one that opens 4,000 connections at once and goes HURRRP [16:06] aria2 [16:06] People too slow. [16:07] i've already hit my limit of knowledge here. Genre style URLs look like http://itunes.apple.com/us/genre/music-r-b-soul/id15 , spit out some XML about jingleDocTypes and jingleActions [16:10] while not the most articulate profile, here's some examples of reviews, comments https://c.itunes.apple.com/us/profile/id1287692619 [16:31] this has not become public since last night: http://archive.org/details/dltv_011_episode [16:31] really [16:32] this item is not yet public [16:32] but its been 12 hours since a post all the info it needed [16:33] mail to archive.org [16:33] when i do the others it up right a way [16:33] mailing to archive.org is like using snail mail [16:34] no it isnt [16:34] they are super nice, responsive and helpful [16:34] i mailed about this problem: http://archive.org/details/abbys_road [16:35] nothing has be done for over a month now [16:35] what problem? [16:35] the stupid meta tags on the mp3 are screwed up [16:36] http://ia601200.us.archive.org/27/items/abbys_road/abbys_road_files.xml [16:36] i can't edit the freaking meta tags to fix it without it going back to normal [16:37] same problem here: http://archive.org/details/this_week_in_fun [16:42] poke underscor [16:46] The item you are trying to edit cannot be retrieved from a "primary US node". [16:47] thanks for getting abbys_road almost fixed [16:47] i still can't edit the items yet [16:48] this week in fun problem: http://ia601207.us.archive.org/33/items/this_week_in_fun/this_week_in_fun_files.xml [16:50] I had that primary US node thing [16:50] usually clears in a few hours [16:54] ok [17:19] metadata that comes from the mp3s needs to be fixed by fixing the mp3s, I think (don't quote me), since I think it would be regenerated from the mp3s every time the item is re-derived [17:26] i fear the possibility of having to back up sites like bandcamp or soundclown er... soundcloud [17:26] but not as much as youtube [17:27] soundcloud would be a great target [17:28] someone elsewhere pasted a link to something on bandcamp that made me consider such sites [17:29] i suppose bandcamp is a selling site, though [17:30] but can have interesting "remix culture" artifacts on it [17:30] such as http://inversephase.bandcamp.com/album/pretty-eight-machine [17:30] and I guess this should go to -bs [17:31] also http://www.audiotool.com/ which i heard of the other day [17:39] Priority-wise, I am very concerned about long-term-existing sites of user data, especially photos and audio/movies [17:39] Hence Picplz [17:40] The music sites that are, basically stores, are another thing. [17:40] Like, I'd be BOTHERED but I feel like, for example, cafepress would be grey [17:42] 5.0G . [17:42] 8.0K . [17:42] root@teamarchive-1:/2/MAILDUMP# du -sh . [17:42] Done three minutes from each other. [17:42] I approve. [17:42] 6.0G . [17:42] root@teamarchive-1:/2/MAILDUMP# du -sh . [17:42] Excellent. [17:50] grr >_< [17:50] note to self: you will be waiting a long time if you don't give grep any filenames to operate on (and no stdin either) [17:52] I do that all the time [17:52] In a similar vein, I was watching a family guy episode and hit pause [17:52] And I thought they were really teasing out the joke [17:52] For 3 minutes [17:53] Then I thought this was either brilliant or insane [17:53] it was pause [18:04] :D [18:04] <3 you because you share stuff like that. [18:05] * SmileyG regularly emails himself, feels his phone vibrate, gets it out to check his new emails only to be disappointed. [18:08] My #1 pet peeve? [18:08] besides underscor of course [18:08] <3 [18:17] godane: Got your redrows deleted. You need to go to the metadata editor and fix the text [18:17] There were invalid characters. [18:17] /var/tmp/autoclean/dltv_011_episode_tmp.xml:15: parser error : PCDATA invalid Char value 28 [18:17] /var/tmp/autoclean/dltv_011_episode_tmp.xml:15: parser error : PCDATA invalid Char value 29 [18:17] [ PDT: 2012-06-14 18:53:30 ] Executing: /usr/bin/xmllint --format '/var/tmp/autoclean/dltv_011_episode_tmp.xml' > '/var/tmp/autoclean/dltv_011_episode_fmt.xml' [18:17] ^ [18:17] architecture, crippled pipelines. Jason Cross of ExtremeTech explains why it's [18:17] ines. Jason Cross of ExtremeTech explains why it's the bargain card that isn't. [18:17] ^ [18:17] ??? [18:17] oh [18:17] I thought that was a mispaste, nvm, it's actually indicating an error in the text :P [18:18] http://hastebin.com/gumiriqaqa.dos is the full log, godane [19:22] thanks underscor [19:23] i just may have retype all of the info instead of copying and pasteing [19:25] underscor: that didn't fix it [19:25] it only fixed for like 5 secs [19:26] :-( [19:31] lzma compression just blows my mind [19:31] 4GB down to under 50mb [19:36] i should have wgotten forumplanet user pages too [19:36] shame on me [19:37] 163 out of 3,124,063 users have been online in the past 60 minutes. [19:38] ok, they are usually empty. nothing too important [19:38] i am thinking of downloading avatar pics though [19:38] ~34k [19:59] instence: lzma is rad. [20:20] almost forgot to redo forums i grabbed without warcs [20:20] 3 days to go :D [20:21] yea i've been doing rar+recovery record, and then an extra backup in 7z+lzma.... because why not? [20:21] yes! why not! [21:14] Anyone wants to help with a little test run of the new seesaw/warrior system? [21:14] There's nothing to actually archive right now, unfortunately, but there is a small example script. [21:14] https://github.com/ArchiveTeam/seesaw-kit#how-to-try-it-out [21:39] godane: Derive is running [21:39] Give it some time [21:39] http://archive.org/catalog.php?history=1&identifier=dltv_011_episode [21:39] There was a long derive backlog [21:41] alard: does that only listen on localhost? [21:41] I suppose I could figure that out by trying it [22:05] underscor: No idea. [22:05] ugh, pip is being annoying [22:07] I think it also listens on other IPs. [22:07] The tornado default. [22:09] i see it now [22:09] its public [23:24] goekesmi: Finished. [23:25] oopa [23:25] oops* [23:25] godane: your item's finished [23:25] not that you're here [23:29] heh. [23:41] SketchCow: I've got a 14GB tar file to push somewhere, containing a mirror of ftp.abit.com.tw as of about January 4, 2009 (as near as I can tell) [23:41] containing bios releases and drivers for a large number of their products. [23:43] I guess I can push it directly, once I determine a suitable item name [23:43] (I made the mirror shortly after the annoucement that abit was closing, though the ftp server stuck around for awhile after that. it now appears to be gone) [23:44] archiveteam-mirror-ftp.abit.com.tw [23:44] err [23:44] godane: your item's finished [23:44] archiveteam-ftp.abit.com.tw or some variant [23:46] are fullstops allowed in item names? [23:46] meh. I'll just s/\./-/g [23:46] many characters are allowed but not really encouraged [23:47] I prefer . -> _ ; it's visually similar [23:47] but yes, I've seen items with . [23:48] . won't be denied [23:48] on the backend, I think it's just anything that works as a unix filename [23:48] but it breaks osme things [23:48] some* [23:48] but they encourage you to stay with [-_A-Za-z0-9] [23:48] we really try for a-z0-9_- [23:48] but capitals are allowed too [23:49] right, and of course please make items not distinguished only by capitalization