[01:01] yipdw: wow. that article shows a gross misunderstanding as to how that bill was created [01:08] oh [01:08] i just noticed who the author was [03:38] i found out that there was a stripper at defcon 4 [03:38] i'm surprise this was defcon.org though [03:53] as part of one of the kickstarter campaigns I funded, they offered to make a game from an idea you had (very quick & dirty), so I put down saving web pages from the yahoo monster [04:34] 13⣎4⣱ 8⡀⣀ 9⢀11⣀ 12⣇⡀ 13⠄ 4⡀8⢀ 9⢀⡀ 11⢹12⠁ 13⢀⡀ 4⢀8⣀ 9⣀⣀ [04:34] 4⠇8⠸ 9⠏ 11⠣12⠤ 13⠇⠸ 4⠇ 8⠱9⠃ 11⠣⠭ 12⠸ 4⠣⠭ 8⠣9⠼ 11⠇⠇12⠇ [04:35] that needs to scroll right-to-left [04:36] 13A4r8c9h11i12v13e4T8e9a11m [04:37] fabulous [04:46] chronomex: op me por favor <3 [04:50] Hey so I'm interested in helping out with archiveteam, is there anything I can do? I have some experience writing web scrapers (python) if that is needed. [04:52] for starters you can run the script at http://archiveteam.org/index.php?title=MobileMe [04:55] >_> [04:55] ok, how much space should I allocate for it? [04:56] I think around 10-20 gigs per profile is fine [04:57] seesaw.sh uploads and deletes after every successful download, so you shouldn't have a lot of accumulation [05:00] ok running :) [05:14] :D [05:17] maybe put that at the end or something :P [05:18] 4,5o12,1h9,6 10,4g5,11o11,13d11,6 5,11m9,1y7,13 4,3e2,6y8,1e12,10s [05:19] I demand a useful topic [05:22] yipdw: Set up UTF8 support, mmkay? [05:25] nitro2k01: I have UTF-8 support on [05:25] that does not make the topic content useful [05:25] well, it is now [05:25] or, put another way, being able to see "SketchCow for President" does not make the topic content useful [05:25] :P [05:27] :D [07:19] DFJustin, are you demanding a +c ? [07:20] am I who's it what [07:22] oh banning colour codes or something? [07:23] nah I don't care [07:24] tpb magnet link archive guy posted his script: http://pastebin.com/8RXXthXB [07:24] "I went trough all the torrents, but as a logged out user. This may be the reason why I saw only 1.5m of torrents. [07:24] I went trough the torrent description sites like [07:24] It took the computer about 2 days. I can post the script somewhere if you want, it is in perl, it is not complicated at all. [07:24] by adding 1 to the $i [07:24] http://thepiratebay.se/torrent/$i [07:25] " [07:25] for reference: from the comments of https://thepiratebay.se/torrent/7016365/The_whole_Pirate_Bay_magnet_archive [07:25] doing it again but logged in would be the first thing i'd think of [07:25] ugh perl [07:25] kennethre: yeah :( [07:25] readable and short enough to easily rewrite though [07:26] yeah [07:26] i wonder how fast i could distribute that [07:26] DFJustin, I thought those colours were a form of protest [07:27] run the whole thing in 10 minutes or something ;) [07:28] kennethre, I think you could download the entire mobileme in a few days :) [07:29] how much disk space do you have? :D [07:29] Nemo_bis: hah, ephemeral so depends on where i'm uploading to ;) [07:30] kennethre, how fast are you able to upload to batcave, usually? [07:30] or to archive.org [07:30] i don't have metrics [07:30] ah, I see underscor is downloading again [07:30] ok [07:30] yep [07:31] Got 2 fresh 2TB drives [07:31] :D [07:31] Plus another 500GB [07:31] i got 180 GB uploaded in <20 hours (iirc) [07:31] underscor, ah, so you're not using the IA machine? [07:31] Liability [07:32] oic [07:32] If it's downloaded from IA IP space, then there are more legal options [07:32] since IA is a "company"/library [07:32] also, they like to keep an arms-length from us [07:32] :D [07:32] since we're kinda crazy and maniacal [07:33] and uploading from the IA machines is not a problem? [07:33] what's being downloaded now? [07:33] mobileme [07:33] Nemo_bis: where's the script? [07:33] Nemo_bis: I dunno, I just want to play it safe [07:33] I don't have a lot of space on my VM anyway [07:34] kennethre, http://www.archiveteam.org/index.php?title=MobileMe [07:34] https://github.com/ArchiveTeam/mobileme-grab/ [07:34] underscor, add it to the topic [07:34] (the wiki page url) [07:34] aww, /this/ is useful topic now [07:36] sweet [07:37] i guess i can spin up a few vms :) [07:37] can i use the same upload slots i used for splinder? [07:37] No, there's a different one [07:37] see the upload-finished.sh script [07:38] i'll need a slot immediately [07:38] Use that, otherwise completion won't get reported to the tracker [07:38] You don't need a slot provisioned [07:38] excellent [07:38] so, we need a +77 % speed increase of the project to complete it [07:38] Really?? [07:39] well then [07:39] i'm on it :) [07:39] 200 instances should do [07:39] kennethre: heroku again? [07:39] :P [07:39] underscor, yes and it was way worse before you restarted [07:39] underscor: but of course : [07:39] seesaw supposedly uploads and deletes completed profiles while downloading or something [07:39] (I haven't looked at it) [07:39] kennethre: These are between a few k and 20GB, so hopefully that's not an issue [07:40] well i can't run a job for more than 24 hours [07:40] one in ...? [07:40] *once [07:41] so a few ginormous ones might need to be re-downloaded [07:41] You don't think heroku will boot you off? [07:41] underscor: i work for heroku [07:41] oh [07:41] :) [07:41] hahahahahahaha [07:42] hehe [07:42] niiice [07:42] how long has this been going on? [07:42] just noticed that the splinder readme had mobileme stuff in it [07:42] "Hey, uh, our bandwidth usage spiked up to 10 gigabits last night. You don't know anything about it, do you?" [07:43] Uh, long time ago [07:43] Before Halloween [07:43] But we didn't actively start until recently [07:43] cf taiwan floods [07:43] thailand floods? [07:44] er, whoops [07:44] thanks [07:45] yeah, seesaw.sh downloads one user, uploads it to batcave, deletes it locally, notifies the tracker, and starts over [07:45] oh geeze [07:45] that's perfect [07:45] hmm [07:46] https://github.com/kennethreitz/heroku-splinder/blob/master/run.py [07:46] so here's what i did before [07:46] https://github.com/kennethreitz/heroku-splinder/blob/master/Makefile [07:46] i'll prob keep using streamer [07:47] Why do that over seesaw? [07:47] different upload location, btw [07:47] (also it's called dld-client now) [07:48] i hope the one drive with that rsync module doesn't get overloaded [07:49] because seesaw only does one at a time, right? [07:49] no concurrency [07:49] so when would your system actually upload anything? [07:50] every 2 minutes it ran the upload script [07:50] in the background [07:50] none of the scripts do concurrency right now [07:50] dld-client nor seesaw [07:50] You just run multiples [07:50] dld-streamer.sh did ;( [07:50] A single thread will max out a 100mbit connection though [07:50] ah they're nice and fast compared to splinder then [07:50] It's akamai [07:51] splinder was all kinds of busted [07:51] shouldn't be an issue then :) [07:51] damn that's a ton of data [07:52] yeah [07:52] 211TB [07:52] estimated [07:52] :D [07:52] seesaw is awesome [07:52] that saves me so much trouble [07:52] anyone doing anything for imgur.com? lots of stuff uses that for hosting. [07:52] splinder was lots of tiny high-latency requests [07:53] yeah [07:53] i wasn't aware mobileme sites were on akamai [07:53] i know the real mobileme had a ton of issues [07:53] i wonder if the data i'm pulling down from my s3 bucket has been marked done [07:53] We should do an imgur mirror [07:54] https://www.youtube.com/watch?v=9MkbLQWHcXw [07:54] but they delete [07:54] :| [07:54] underscor: i bet there's a ton of illegal content [07:54] and i forget if i uploaded it to the batcave. (i suspect i did) [07:54] kennethre: Archive first, (don't) delete later ;) [07:54] haha [07:55] Yeah basically :> [07:56] wow and get-wget-warc too [07:56] this is way better [07:56] easy as pie [07:56] is there a verbose mode for any of this stuff? [07:56] i pretty much can run and look at logs [07:56] Does heroku give you basically a VM or something? [07:56] underscor: kinda, abstracted away [07:57] DYNOS! [07:57] :P [07:57] yipdw: :) [07:57] also yes there is a ton of illegal content on mobileme [07:57] I must have like gigs of new music [07:57] oh imeant imgur [07:57] most of which I am sure does not come with rights for redistribution [07:58] ohh [07:58] yeah that too :P [07:58] I must have like gigs of new music [07:58] hahaha [07:58] interesting project would be to pull out all the music we find [07:58] I haven't actually checked all the WARCs [07:58] or movies [07:58] but some of the biggest ones I've found contain either that or QuickTime movies [07:58] I'm sure there's lots of movies too [07:59] home movies most likely [07:59] so ./seesaw.sh just exits when a user is done [07:59] oh, it uploads the user's data, deletes it, goes for another [07:59] so autonomous, excellent [08:00] i assume it repeats and repeats [08:00] yeah [08:00] it's too bad MobileMe is not hosted on e.g. S3 [08:01] Why? [08:01] because then Heroku dynos would have a hilarious speed advantage [08:01] yeah [08:01] (afaik they're on EC2 instances) [08:01] and free [08:01] heh [08:01] oh, hahah [08:01] transfer [08:02] we'll get crazy speed anyway [08:02] i feel like upload will be the bottleneck [08:02] :P [08:02] I look forward to seeing kennethre bogart the dashboard [08:03] i really hope it happens :) [08:03] not fair >:| [08:03] I've been trying hard to get to the top spot [08:03] i'm not getting my hopes up [08:05] alright, here goes nothing. [08:05] how many processes? [08:05] haven't turned it on yet :) [08:05] oh [08:05] alright, 1 [08:05] ONE [08:06] gah failure [08:06] gah it relies on wget [08:10] That's what get-wget-warc does [08:10] Oh wait, does get-wget-warc call wget? [08:10] haha [08:10] yeah it does [08:10] i swapped it out with curl [08:10] alright i'm in action [08:11] i need to write an archive team buildpack [08:11] that has warc prebuilt [08:11] and rsync [08:11] the next wget release will have WARC writing capabilities present, but I'm not sure when that release will be [08:11] kennethre, yes, I doubt you'll be able to upload everything immediately [08:11] batcave is too slow for that [08:12] it's fast here [08:12] and it doesn't have a lot of space [08:12] I regularly get at least 1 MB/s [08:12] yipdw, how fast? [08:12] /dev/vdc 16T 15T 1.5T 91% /2 [08:12] uh oh [08:12] that's slow :-p [08:12] D: [08:12] it's not too bad [08:12] (that's the disk mobileme uploads are going to) [08:12] hmm [08:12] maybe those should be cleared off [08:12] it is if you download at some 10-100 MB/s as kennethre probably will :p [08:13] it's not too bad so far [08:13] he has so far downloaded < 1 MB [08:13] :P [08:13] haha [08:13] building now [08:13] underscor: where exactly is the mobileme data going to end up? [08:13] 200 TB is quite a tall order to provision [08:14] SketchCow is talking with archive folks [08:14] I think the current plan is to stick things in temporary items on the cluster [08:14] then slowly "download" them and reorganize and then reupload [08:14] Where's the dash? [08:15] Since they don't have 200TB of disks just laying around [08:15] hahaha [08:15] ersi: The dash? [08:15] +board [08:15] memac.heroku [08:15] .com [08:16] neat [08:18] If I'm running dld-client.sh, does stats get reported to the tracker? [08:18] yeah [08:19] alright [08:19] i think i'm in action [08:20] hroku scale scrape=100 [08:20] :D [08:21] do you get free resources? [08:21] yeah, my bill is $0 every month [08:21] no matter what i use [08:21] obviously i shouldn't do anything crazy [08:21] hopefully this won't be too intensive [08:23] TThat's cool [08:24] wow it's flying [08:24] this worries me [08:24] in the right on the dashboard [08:25] those are already-uploaded, right? [08:25] No, those are completed [08:26] They don't get reshown when they upload [08:26] Are you using seesaw? [08:26] i am [08:26] you weren't kidding when you said it'd exhaust the connection [08:26] damn [08:26] aaand there it goes [08:27] ho' shi', boy! [08:27] that's crazy [08:27] I want to see cacti/mrtg graphs for the ec2 instances [08:27] :D [08:27] hopefully this will help you keep up :) [08:27] this kind of saturation worries me [08:27] i need to check a few things [08:27] haha [08:27] i wasn't expecting it to go nearly this fast [08:28] "Uh, Kenneth. You broke EVERYTHING!" [08:28] nah not at all [08:28] haha [08:28] just abusing resources ;) [08:28] "slightly" [08:29] can someone tell me how much my user has uploaded so far? [08:29] is there an api for that? [08:29] just trying to gauge the current speed [08:29] kennethre: I can, one sec [08:29] No API, but I can du your module [08:30] it's been up for ~8 minutes [08:31] du: cannot read directory `kenneth/m': Permission denied [08:31] Oh, nevermind then [08:31] :( [08:31] I don't have sudo rights on SketchCow's box [08:32] kennethre: Upload your files with global read and exec [08:32] :P [08:32] Everyone elses are 755 [08:32] underscor: hey i'm just running the script :) [08:32] Yours are 700 :P [08:33] well a bit late for that now ;) [08:33] Heroku bug? [08:33] Well [08:33] Not bug [08:33] but default [08:33] prob just a configuration difference [08:33] yeah [08:34] Hmm [08:34] Oh, hey, mobileme isn't on /2 [08:34] Whoops [08:34] haha [08:34] /dev/vdd 11T 8.4T 2.2T 80% /3 [08:34] So, little more breathing room [08:34] if you're curious, here's what i'm doing https://github.com/kennethreitz/heroku-mobileme [08:35] Wow, you're killing the dashboard [08:37] :) [08:37] this is far more fun than you think. [08:37] i could scale it up way higher [08:37] but i think this'll do for now [08:38] underscor: probably another umask for his user, either on the upload box or download boxes [08:39] kennethre: /me waits for akamai to call up heroku [08:39] "Hey, uh, could you stop ddosing us?" [08:39] :D [08:39] that would deeply suck [08:39] i think they'd call ec2 first [08:39] I bet Apple will get a nice bill from Akamai [08:39] hehe [08:39] I'm sure that they're used to much higher load [08:39] Monetary Denial of Service [08:39] HAH [08:40] yeah, I bet they see hueg spikes occationally [08:40] have you seen their room? [08:40] Batcave is receiving at 10MB/s [08:40] but yeah, I bet kennethre is throwing them a punch haha [08:40] (so 80mbps) [08:40] http://www.youtube.com/watch?v=kiZgsBUEJdA [08:40] (Inside Akamai's war room) [08:41] i think it's pretty obvious we're not ddosing [08:41] Oh man [08:41] That's cool! [08:41] I want to visit that room [08:41] it's not like we're slow lorising them :) [08:41] hehe [08:41] oh man, youtube commercials for the fail [08:41] Oh, wow [08:42] It's Ahhh kamai [08:42] not aaaaa kamai [08:42] I've been saying it wrong for a while [08:42] start spelling it out instead, a k a m a i [08:43] can't go wrong with that [08:43] :D [08:44] Is anyone else pausing it so you can read everything on each display? [08:44] haha [08:44] http://www.youtube.com/watch?feature=player_detailpage&v=kiZgsBUEJdA#t=26s [08:44] 2.6 terabits/sec?!?!?!?! [08:45] hahaha [08:45] That's SO COOL [08:45] they power like 70% of the internet [08:45] haha [08:45] 12.7 million requests per second [08:45] wow. [08:46] kennethre: Batcave intake has jumped from ~50-80mbps to ~184-220mbps since you started [08:46] kennethre: https://github.com/kennethreitz/heroku-mobileme/commit/3f8ce58702cdbdf5dca3f48e4aa112b111d2a3a1 <-- hah what [08:46] yipdw: yeah, i'm just telling heroku it's a python app [08:47] yipdw: since i'm too lazy to do a custom buildpack [08:47] heh [08:47] underscor: that's freaking awesome ;) [08:47] haha [08:47] I wish I had sudo [08:47] underscor: what kind of connection does that thing have? [08:47] Then I could run iftop [08:48] Heh, first user I got; - Running wget --mirror (at least 16520 files)... [08:48] 400MB downloaded atm [08:48] kennethre: Gigabit copper [08:49] underscor: strait to a backbone? :P [08:49] No [08:49] i'm joking [08:49] I have a whole 1 mbps up [08:49] so assuming i can keep this up, we'll hit the deadline? [08:49] VM (batcave) -> host (gbit copper) -> switch -> switch -> router -> router -> internet [08:50] kennethre: http://ia700000.us.archive.org:8088/mrtg/ [08:50] If you want to watch some stats [08:50] ah excellent [08:50] thanks :) [08:50] haha [08:50] I like the spike [08:50] The network graphs are in bytes, fyi [08:50] i'm really tempted to spin up another 100 [08:51] This host is 128GB ram [08:51] * underscor wishes he had that to play with [08:51] kennethre: do eet! [08:51] lol [08:51] err [08:51] is this a good idea [08:51] hmm [08:52] Wow, the disk is filling at about 20MB/s [08:52] 2271450 MB left [08:52] heroku scale scrape2=100 [08:52] 31 hours until the drive is full [08:52] will take about two minutes for 100 machines to build rsync and wget [08:52] :D [08:52] i hope there's another drive [08:52] no, SketchCow will have to do something [08:53] Either handle it himself or give me access [08:53] because he's in finland [08:53] this makes me appreciate the mind-boggling scale of EC2 more [08:53] ah [08:53] documentering? [08:53] yeah [08:53] and attending some hackfest or something [08:53] helsinki hacklab iirc [08:53] awesome [08:54] who wants to take bets as to when kennethre starts getting graphed? [08:54] haha [08:54] 30 hours [08:54] 15 [08:54] well hmm [08:54] maybe i should spin up another 200 [08:54] and try to stay up until it happens [08:54] wtf [08:55] lol [08:55] is that 200 dynos? [08:55] yes [08:55] what the [08:55] I can't even price that on your site [08:55] haha [08:55] $0.05/hour [08:56] per dyno [08:56] oh jeez [08:56] so currently $10 an hour [08:56] considering what i've downloaded so far, that's a hell of a deal [08:56] Alright, if we get a "Heroku is out of business" news from somewhere - we know who to blame [08:56] :P [08:56] haha [08:56] That's actually a really good deal [08:56] well we're currently running 1.1 million apps [08:57] If we ever needed to do a really fast "oh shit" backup [08:57] yeah if that ever happens, call me :) [08:57] lol [08:57] What's stopping someone from using that as a DOS manuver? [08:57] maneuver* [08:57] nothing really [08:57] threat of being fired [08:57] :P [08:57] haha [08:58] You could fire up a bunch of pre-built "attack VMs" and up on any platform [08:58] Rackspace, Linode, EC2 [08:58] I also think we probably DDoSed Splinder [08:58] heroku makes it particularly easy since we're app-level [08:58] yeah, to hell and back [08:58] not system-level [08:58] kennethre: true, but making a pre built image isn't hard :) [08:58] even though I bet it's harder than whatever you did [08:58] ersi: takes longer than 5 minutes [08:58] also, YouTube shutting down in 24 hours [08:58] considering time and details [08:58] lol [08:58] yeah [08:58] yipdw: wat [08:59] but yeah an ami for this would be awesome [08:59] no just kidding [08:59] one of the XXLs [08:59] somehow I don't think even all of EC2 could save YouTube in 24 hours [09:00] haha [09:01] it's possible [09:02] my dad wil soon have access to the storage [09:02] *will [09:02] he's going to go work on a system with 230PB of storage :) [09:02] wow [09:02] he'll be working to get them 1TB/s io [09:02] imaging system? [09:02] (not a typo) [09:02] LHC? [09:02] hah [09:02] let me find it [09:03] kennethre: http://tracker.archive.org/df.html [09:03] that's the disk space remaining on batcave [09:04] O_o [09:04] (live updating) [09:04] title that COUNTDOWN TO DESTRUCTION [09:04] lots of programs tend to go haywire when there's no more space left [09:04] http://www.informationweek.com/news/government/enterprise-architecture/231900554 [09:05] yipdw: refresh [09:05] that's the computer he'll be working on [09:05] underscor: I approve [09:05] kennethre: ooh awesome [09:05] 960 Tesla M2090 GPUs [09:05] :O [09:06] yipdw: that's the old one that's being replaced ;) [09:06] oh wait [09:06] oh wait nvm [09:06] you're right [09:06] oh [09:06] kennethre: woah, that's cool [09:06] yeah, that's phase 1 [09:06] "outfitting the machine with as many as 18,000 Tesla GPUs" [09:06] phase 2 appearently is "outfitting the machine with as many as 18,000 Tesla GPUs" [09:06] hah [09:06] lol [09:06] it's going to be awesome [09:07] I've been learning CUDA programming in my spare time and even the little toys I've run on my GTX 560 Ti blow me away [09:07] ah only 55PB of storage [09:07] scaling something to 18,000 GPUs must either be awesome or hell [09:07] i can't imagine [09:07] i have a hard enough time understanding gevent :) [09:07] I imagine you're going to hit bandwidth issues with main memory <-> GPU memory transfers [09:08] yipdw: There, I added some graphics too [09:08] what the [09:08] man, that is declining fast [09:09] yeah, 26MB/s [09:09] Oh, wait [09:09] 36MB/s now [09:09] kennethre: you're insane :D [09:10] I do what i can :) [09:10] web.me.com/venstrehaandsarbejde is taken over 1GB now >_> [09:10] I should probably email SketchCow [09:10] Subject: NEED NEW DRIVE NOW [09:10] It's 11:11 AM over in Finland right now btw [09:11] e [09:11] so should be awak [09:11] oh, cool [09:11] not on skype though [09:11] Overall Top RX Speed: 118128.65 KB/s [09:11] ^ kennethre [09:11] 118 megabytes!!!!!!! [09:11] oh god [09:12] That's actually maxing the disk controller [09:12] :D [09:12] I just noticed he checked in something like 8 gigabytes of data [09:12] maybe faster than 15 hours [09:12] it's been what, 30 minutes? [09:12] hahaha [09:12] 20 [09:12] 52 for the first half, 20 for the second [09:12] hah [09:12] Something else about mobileme: It's not really urgent, but I've just updated the wget version. If you have the time, please run git clone and get-wget-warc to get the latest version. There was a problem downloading files >2GB to a WARC file, which should be fixed in the new version. (Not that there are many users with such large files.) [09:12] Usage at 12:55AM GMT was 2267348 [09:13] haha [09:13] Er, disk free [09:13] more build times ahead [09:13] So [09:14] 29GB in 19 minutes [09:14] (not all kennethre, but mostly) [09:14] Wait, that doesn't mke sense [09:15] Oh, wait, yeah [09:15] Nevermind [09:15] I was thinking that was 29GB in 19 seconds and going "wait a sec... lol" [09:15] So 1528MB/min [09:15] haha [09:15] i was doing my math at 118 a minute [09:15] not second.. [09:16] 1462 minutes until dday [09:16] alard: So.. files over 2GB will be capped? Is it fixable/noticable? [09:16] So roughly 24 hours at the current rate [09:16] No, they won't be capped. wget will just crash. [09:16] and it's accelerating [09:16] alard: hah, awesome. Well, I guess that's noticable [09:16] Ah, that explains the one user that kept segfaulting [09:17] Yes, that's probably the reason. [09:17] I discovered the problem with some users on umich.edu. [09:17] kennethre: Idling around 49MB/s currently [09:17] weak [09:17] :) [09:18] If that gets close to 0, can you call off your army? [09:18] haha [09:18] unfortunately i'm having some processes being killed [09:18] Aww, why? [09:18] Hm, does doing this with warc mean that we get the data * 2? [09:18] memory quota [09:18] exceeding 512MB [09:18] oh, yeah, they do that [09:18] rare though [09:18] yeah [09:18] usually just really big users [09:19] whoa, wget is using like no memory at all these days :o [09:19] yeah, thank alard [09:19] (I think it was him?) [09:19] alard: you the man [09:19] Or who ever triaged that --mirror bug :) [09:19] there were some things that were not free'd [09:20] yeah, was like two three memleaks or something like that if I recall correctly [09:24] huh [09:24] I might just be crazy but I think a terabyte or two has been downloaded in the past half hour [09:24] from mobileme [09:24] at least I remember the total being 11000-something GB [09:24] and now it's 13036 [09:30] underscor: hahahaha [09:30] underscor: the countdown to destruction [09:34] wait [09:35] efnet has a channel limit????? [09:35] * #retroshare :You have joined too many channels [09:35] yeah, all networks have a channel limit [09:35] oh, hah [09:36] I guess I've just never hit it before [09:36] It's usually 20 [09:36] but could be anything [09:36] 20 is far too low [09:36] jebus, 2GB on this user [09:37] Hurr.. I'm inserting 48*3600*24*31 rows into MySQL [09:42] i really want to spin up more [09:42] i'll wait until the disk problem goes away though [09:43] haha [09:47] kennethre: http://ia700000.us.archive.org:8088/mrtg/iowait.html hehe [09:58] jeepers, what are you doing [10:03] proving that the cloud goes to 11 [10:04] FreeNode doesn't seem to have any limit [10:04] I once joined some hundreds (thousands?) channels [10:05] and how did irssi like that [10:07] not well, I expect [10:07] irssi handles that fine [10:07] I've spammed 4000 PM's at irssi just fine [10:08] ie unique windows [10:08] but 4000 active channels? [10:08] same thing more or less [10:08] mIRC crashes if you do the same thing anyhow [10:09] irssi didn't even notice I was doing something strange [10:09] I am shocked, shocked that mirc doesn't like that [10:09] sent like 500 PMs to a mIRC instace and it b0rked [10:09] you only have to set a timer because there's a flood limit on the network [10:09] Yeah, me neither. But when I was doing what I was doing.. [10:09] yeah [10:09] you'll hit the buffer limit very quickly if you try to join them all at once [10:09] yep [10:10] and also get kicked by some overzealous staff perhaps [10:10] because they have all sorts of smart reporting bots [10:11] think i'm going to bed [10:11] should i spin up more instances first? [10:11] eh i might as well stay up [10:15] I want moar [10:15] kennethre, how's your RTT to batcave? [10:16] good question [10:17] wow, no ping [10:17] easier said than done :) [10:17] rtt min/avg/max/mdev = 192.448/246.892/286.526/27.767 ms [10:17] twice as faster than usual from here, who knows why [10:18] because we're awesome [10:18] and cloud scale [10:19] alright, don't kill me [10:19] and still the traceroute seems more silly than usual [10:19] i'm feeling ballsy tonight [10:19] i'm about to get 50% faster [10:20] then sleeping [10:21] sleep 50% faster? [10:21] sounds good to me [10:21] I'd like to sleep less [10:22] me too [10:22] i meant downloads [10:22] :P [10:32] BAM [10:32] Morning [10:32] i made dashboard [10:32] kennethre: I assume you're slamming the moblieme download. [10:33] SketchCow: indeed — 387GB in 2 hours ;) [10:33] Where's the dashboard? [10:33] http://memac.heroku.com/ [10:36] 26 days to complete it at this speed (theoric) [10:41] Is the system you're using marking things as "done" so I can throw them into archive.org? [10:41] SketchCow: I was just going to write: 'Let me know if you need help with the processing of the mobileme rsync uploads.' [10:42] There's a list of users that have been uploaded to the new mobileme rsync module. [10:43] If that's useful, I could write you a little script that monitors that log and moves users from the upload directory to another place when they're done. Maybe making batches of 100GB each, which you can then move somewhere else when you're ready. [10:43] Or I could just give you a pointer to the log. [10:48] I need help, let's put it that way. [10:49] Because if we really are slamming in 387gb in two hours, I really need to start pushing that stuff out to archive.org proper ASAP. [10:49] A script written by the very guy who wrote the script that is putting in the data in the first place seems the superior solution. [10:51] haha [10:51] What happens is this: 1. The upload script rsyncs an user to batcave/mobileme/UPLOADER/U/US/USE/USER/ [10:51] 2. The upload script sends the message 'USER uploaded by UPLOADER' to the tracker. [10:52] Now, somehow, there should be something that monitors that log and processes the uploaded users, moves them from the rsync path to somewhere else. [10:52] Yes. [10:53] Do something that is run from mobileme. [10:53] It's OK to write a hook and let me write the final uploader. [10:53] In many cases, we're probably going to .zip up the user [10:54] Do you want one big stream of individual users? Or do you want batches? [11:01] Let's talk about this. [11:01] I keep getting pulled into things, so I might disappear, but... [11:02] I need to add sets of items to archive.org. It would be better for us to pop together 200gb sets of persons into archive.org's sets. [11:02] You know, roughly 200gb. Could be 150gb if needed, say. But don't want to add billions of items into their DB. [11:04] SketchCow: do you cover http://en.wikipedia.org/wiki/UltraCade_Technologies in ARCADE? [11:04] My schedule. I'm in Finland today, at a newspaper, actually, watching my fiancee' talk to 50 reporters, management and others about the future of content. [11:05] Schbirid: Wrong channel [11:05] oops [11:05] #jsmess ? ;D [11:05] I'm online, and I can talk about archiveteam stuff but I suspect we'll be onto a new archive. [11:05] Just e-mail me jason@textfiles.com with "I herd u makin movie plz include xxxxx thnks" like the 4,000 other people have. [11:06] heh [11:06] maybe! [11:06] Let's try my last line again. [11:06] i am researching the quake arcade tournament cabinet at the moment [11:06] channel as in communications channel, not IRC channel [11:06] aye [11:06] i am german, is that excuse enough? [11:06] "I'm online, and I can tak for now, but I am accompanying her tour today, and so I might be puled away." [11:07] Gesundheit [11:07] I owe one of my jobs an article tonight, that will take my attention. [11:07] MobileMe feels darn large, and I've only done 10-30GB as of yet :/ [11:08] But after this, I can really concntrate on the data move, so alard, that would help a lot. [11:08] i kinda want to try archiving http://home.swipnet.se/ [11:08] Also, batcave is going away, moving to a new machine. [11:08] seems straightforward, it is URLs like ~w-25876/ [11:08] err http://home.swipnet.se/~w-25876/ [11:08] I want to move stuff into archive.org rather than the new machine. [11:08] and you just increment the number [11:08] So that's the news. [11:08] not sure if the character too [11:09] ah, google says there are normal urls too, as in /somestring/ [11:09] Schbirid: I'd be on for that, considering being a Swede and all [11:09] SketchCow: Let's move this discussion to #memac [11:10] Schbirid: I'd also like to fetch all of http://medlem.spray.se/member/ [11:10] one of the old "geocities of Sweden" [11:10] :) [11:10] alard: Excellent idea [11:10] ersi: Is it in danger of disappearing? [11:11] Well, it's owned by a bunch of morons [11:11] so they're of the flippity flop kind of people [11:11] but they've managed to stay online as of yet, dunno if all of their legacy material is though [11:12] I'll start looking into it though :) [11:13] it is amazing how information dense those "oldschool websites" often are [11:13] not some polished empty fluff [11:13] mostly just text/images/sounds, ie content :) [11:14] yeah [11:16] "If you cannot see the navigation window on the left, or the animation on top, please download Macromedia's Flash plugin." [11:35] I am all for "warning", is all [11:35] But we are downloading a LOT of data right now. [11:35] Mobile Me... I mean, ow, that thing is huge. [11:44] Yeah, of course. That comes first [11:44] Doesn't hurt to scout around and plan a little though :) [12:06] Back one a gain. [12:17] Now I'm in this beautiful meeting and symphony/performing space in Lahti, Finland. [12:51] sweet [13:03] kennethre: Drop to half of what you're doing. [13:03] You're killing that machine and at current rates I will not have the push out to archive.org done before you've filled the machine. :) [13:04] We're too good at acquiring data and I need to do one more round of pushing files out. [13:07] I think he might have gone to sleep [13:08] He was considering it two and a half hours ago at least [13:15] YEah [13:15] We're gonna get raped [13:15] I see where this is going [13:15] The machine is actually slowing down and not working well now. [13:15] You could shut down the tracker temporarily [13:15] that'll make him stop fetching jobs [13:15] I could shoot the dog, that'll make him stop eating [13:16] Indeed [13:16] I was merely suggesting that maybe it's better to die prematurely with disk space left, than having everything go until full stop crash and burn [13:16] but why does network decrease and CPU increase at the same time [13:26] Nemo_bis: Network functionality isn't free :) If the CPU is spending all its time going "oh shit writewritewrite" it's going to start ignoring some packets. [13:27] sure, I just don't understand the chronology [13:58] SketchCow: cut in half [14:04] 1/3rd [14:05] Nice, kennethre is now #4 on the dashboard [14:05] I was hoping to be #1 today :) [14:05] show off :-) [14:10] :) [14:11] I'm just sad the dash doesn't show more users :( [16:47] underscor, anyone else who was watching: so how did it take kennethre to get in the top 10 on memac :P [16:47] er, how long [16:48] yipdw|m: 2 hours :) [16:48] jeez [16:48] I had to scale back, batcave wasn't doing well [16:48] oh, I see that from the dashboard charts nw [16:48] now [16:49] if you keep the charts at their default scale it's just a vertical line [16:49] hahaha [16:52] hehe [16:52] the graphs should automatically switch to log scale as soon as kennethre appears [16:54] ha [17:41] man [17:41] CTA trains today are a mess [17:41] seems there was a derailment [17:42] threw off the entire Red Line [20:06] oh bollocks, i did not check my robots.txt downloader for weeks and it did not download... [20:07] wait, false alarm [20:08] nice, 212 days of files [20:21] find anything interesting? [20:22] me? i havent poked around in it for months but when i started it was quite fun [20:22] i should find a solution for sharing them [20:27] "Boob is the only word in English which contains full 3D view of the object. Top view: B. Front View: oo. Side view : b" [20:33] Hello everyone. :-) [20:34] I have a question about mobileme-grab. [20:35] wget-warc has been updated to a new version that allows downloading of bigger files, but should I worry about the data that I've downloaded previously with the old version? [20:36] ask away, eprillios egfault [20:36] whoops [20:36] What would happen before is wget would segfault [20:36] So you'll know if you had any with issues [20:36] They wouldn't get marked as "done" anyway, so it's cool [20:37] Okay, thank you. :-) [20:37] no :) [20:37] np* [20:37] Thanks for helping out!