[00:01] I feel like I'm watching a film Mr. McFeely delivered to Mister Rogers [00:01] well, until the automation sets in. that's a bit different [00:03] i rather like it :) [00:03] then it goes a bit wonka [00:17] Won the 1959 doc oscar [00:20] oh, i'm not imagining having seen it before then [00:47] http://archive.org/search.php?query=collection%3Aafterhoursdjs_livesets&sort=-publicdate [00:47] SO MUCH TRANCE [00:57] aw mai gawd [01:31] oh wow, wasn't even recognizing that. i love ah.fm [01:33] di.fm, ah.fm and pure.fm [01:33] are all really good [03:27] Ha ha arguing with google guy about death of google reader [03:30] lol [03:30] I still haven't moved off Google Reader [03:30] I really should [03:37] SketchCow: rofl [03:40] "of course there was a reason. you just may not agree with it." [03:40] what sort of bullshit fucking reason is that [03:44] "hi, I'm hitler, there was a reason I killed millions of jews in the holocaust, you just may not agree with it" [03:47] "You must be a real hoot on the slave ship." [03:48] oh my god that was awesome [03:55] but i love how he picked on the literal meaning of "no reason" as if you actually meant literally no reason, as if they just tossed a fucking coin with "kill google reader" on one side and "eat a burger" on the other [03:55] rather than the obvious meaning of "no non-bullshit reason" [03:55] i *love* it when people go selective-sperging like that [03:57] It is just a deflection technique to try and discredit an argument which itself shows how weak the other position is [04:10] yes [04:11] in related news, there are still people who will get into a twitter fight with jason [04:11] aka "bunny meats blender" [04:11] not gonna work out too well for fluffy [04:13] I always thought of Jason on twitter like a giant meteor on the way to destroy your planet. You can say anything you want but it is not going to change the outcome [04:14] kinda like this http://medias.omgif.net/wp-content/uploads/2011/08/Real-bullet-bill-real-attack.gif [04:14] I get the image of Jason in a tuxedo slamming head first into Yahoo HQ [04:18] ehonda style? [04:20] oddly, reader is still up for me [04:23] Still loads for me, though I've moved everything to newsblur, which is FOSS so I can host myself even if the main site goes down. [04:46] what google guy? [05:09] Seth L. in this thread https://twitter.com/textfiles/status/351866764289769472 guy just doesn't get it [05:20] ah, ty [06:32] I've added a new tool. [06:32] http://www.archiveteam.org/index.php?title=User:Jscott/Sorry_That_I_am_All_Up_In_Your_Shit [06:32] So, in the future, feel free to link someone I'm arguing with [06:35] hahaha [06:37] oh my god haha [06:42] very good [06:56] SketchCow: at some point your going to have setup a theblazetv-hightlights collection [06:56] i only say that cause grabing all of them [06:57] good news is my brute force xml grab is working [06:58] getting lots of bad data but just have search for any file with 'Page Not Found' and remove it to fix that problem [07:11] godane: all of that will get the job done, though if you're not in a rush for a current project, it might be good to try to work some python out [07:36] https://www.google.com/reader/about/ [07:36] API is still up [07:41] API is down [07:42] RIP Google Reader [07:42] whwhwhatt [07:43] no [07:43] getting lots of errors [07:43] :( [07:44] curse you mihaip and your memory leaking totally free tool [07:55] and that's my migration to RSSOwl done [08:18] aw [08:18] reader has finally died [08:26] * winr4r salutes ivan` [08:49] winr4r: you can also thank alard for writing most of the necessary software beforehand [09:03] I think the ArchiveTeam wiki is getting HNed, though it still responds after a while [10:03] https://news.ycombinator.com/item?id=5976263 Google Reader is dead (google.com) 134 points by voidfiles 2 hours ago | 79 comments [10:14] . [12:36] * Baljem hrms and hopes SketchCow never has reason to send him that link [12:37] as generally I fall into the first category of boring people, who just happened to notice far too many instances of Jason being right so ended up getting involved ;) [12:47] Hackaday being possibly sold/moved on.... [12:49] :O [12:49] O_O [12:49] source? [12:55] http://hackaday.com/2013/07/01/hackaday-looking-for-a-good-home/?utm_source=feedburner&utm_medium=feed&utm_campaign=Feed%3A+hackaday%2FLgoM+%28Hack+a+Day%29 [12:56] ah [12:57] luckly i did a backup of hackaday last year [12:58] and its right here: https://archive.org/details/hackaday-2004-2011-20120730-mirror [13:15] pushing my mirror of www.boilingfrogspost.com [13:15] this has no mp3 in it [13:34] uploaded: https://archive.org/details/www.boilingfrogspost.com-20130623 [14:31] Do another backup of hackaday, please. [14:42] for somebody with access to the IA collection, could my google answers archive I made a long time ago get moved [14:43] http://archive.org/details/google-answers-archive to the Archive Team collections, IIRC. I can't do it, someone with IA edit access will have to. [14:45] yeah that [14:52] Done. [15:04] thanks to me mirroring techcrunch [15:05] i may have found the right way to hackaday.com [15:05] wget "$website/$year/" --mirror --warc-file=$website-$year-$(date +%Y%m%d) --warc-cdx --accept-regex="(/$year/|/common/images/|\.jpg|\.png|\.jpeg|\.gif)" --reject-regex='(\?)' -E -H --domains=$website,weblogsinc.com,files.wordpress.com -o wget.log [15:09] http://greader-items.dyn.ludios.net:32047/common_crawl_index_urls.bz2 in case anyone wants a 22GB dump of commoncrawl URLs, will be down in a day [15:09] what's the best way to mirror phpbb forums? [15:09] you can get the same thing from the common_crawl_index tool but it takes two days to download uncompressed from a public S3 dataset [15:20] fixed: wget "$website/$year/" --mirror --warc-file=$website-$year-$(date +%Y%m%d) --warc-cdx --accept-regex="(/$year/|/common/images/|\.jpg|\.png|\.jpeg|\.gif)" --reject-regex='(replytocom)' -E -H --domains=$website,weblogsinc.com,files.wordpress.com -o wget.log [15:20] I can grab that, ivan. [15:20] Are we sure Archive.org doesn't already have that? [15:21] i may have just need to block replytocom and not just anything with ? in the url [15:21] http://archive.org/details/commoncrawl [15:21] hmmm. [15:22] 5.93M/s eta 1h 43m [15:24] 3.92M/s eta 39m 24s [15:24] Not bad. [15:24] So, people will need to keep reminding others of this, but right now only underscor and I have access to the IA Archive Team collection. [15:24] I can swap things over quickly. [15:25] Unfortunately, there's no way to grant anyone here the ability to make another person's item enter the archive. [15:31] 2004 hackaday.com urls is downloaded [15:38] balrog: there are scripts to scrape phpbb to other forums. [15:38] I've done DB dumps of a few hosted fora that way. [15:59] thanks for the Google Answers archive move [16:33] hooray [16:33] * winr4r updates the wiki [17:05] so as the reader API is dead, shall i move the "how to get involved" stuff into a /Warroom or /Archive subpage of the google reader page on the wiki? [17:05] Yes. [17:05] okay [17:05] shall i do that for other ended projects as i encounter them too? [18:13] while decommissioning a machine, i came across some files from the yahoo video project that I don't think ever made it to their final resting place with the internet archive. anyone got a pointer on what I might do with them? [18:13] zumthing: ping SketchCow [18:13] and whereby files, i mean 900GB of data [18:14] winr4r: i shall try that again. [18:15] probably quicker to post him the drive or something [18:15] lol, 900 gigabytes [18:23] SketchCow: did you grab the "last 11 posterous blogs" item? [18:24] I seem to recall I'm some sort of admin but I forget what to do right now. [18:44] zumthing: I'll give you an ftp - will that work for you? [18:45] winr4r: thanks for cleaning all that up [18:45] ivan`: any time :) [18:45] SketchCow: that would, or I'm just trying to track down a disk to mail you. [18:46] SketchCow: i found the archive.org site related to the upload, perhaps I can pare down what I have to not duplicate what's already there. [18:47] ivan`: also as you're a wiki admin now, you should probably add xanga to the front page under current projects [19:03] winr4r: added, let me know if it should say anything else [19:13] zumthing: Up to you [19:14] SketchCow: did you grab the files I had, or do you want me to upload them to an ftp? [19:14] I officially forget. [19:15] if you want, just give me an ftp and I'll upload [19:15] this is the pile of google video files [19:16] I nominate winr4r for wiki adminship based on his great edits and superior command of English [19:17] also I need a break and don't want to edit at the moment ;) [19:33] https://archive.org/details/BuyValiumCheapCodNoRxValiumOnlineOvernightDeliveryCodOvernight lol wut? [19:37] We find those and delete those [19:37] Are you just now finding out we have spam to deal with? [19:38] haha yes [19:42] A lot. [19:42] We deal with a lot of spam. [19:49] ivan`: thanks :) [19:50] but seriously I might be MIA for a while and would like the homepage to be less broken ("add link here...") [19:51] IVAN: SHORTEST ADMIN REIGN EVER [19:51] haha [19:52] you're like the lady jane grey of wikis [19:53] Ivan fired, Winr4r now an admin [19:53] thanks [20:08] okay that's Recently Ended Projects updated [20:09] cool [21:19] does anyone want a bzip2'ed version of greader feed stats that is ~25GB instead of ~130GB of .warc.gz? [21:19] also available over fast internet only briefly [21:22] ivan`: not me personally, but is it in the archive.org collection? [21:22] seems it should be! [21:22] the .warc.gz's are [21:22] I guess it could be [21:23] yes [21:34] http://techcrunch.com/2013/07/02/yahoo-acquires-qwiki-for-around-50-million/ [21:36] i wonder how this one's going to end! [21:40] "Thank you for being a part of our story - one which is far from over." [21:40] i swear there's one guy who's hired to write all of these blog entries [21:41] all of the "we got acquired" ones that is [21:45] in any case, this is an obvious talent acquisition, you don't pay $50 million for a fucking slideshow generator because you think it's going to make you more money [23:58] OK, so let's figure out how to download them.