[02:06] Co-worker and I did this today. [02:06] https://archive.org/details/a2_XPS_Diagnostic_v1.05_1983_XPS [02:07] Runs JSMESS headless, generates screenshots, gets uniques, makes an animated gif as screenshot. [02:07] Insta-preview [02:09] sweet [02:09] Still some parameters to tune [02:09] But surprisingly good [02:09] And best of all, I can leave it running for days [02:09] it just sits there doing the work [02:09] We might go back and clean up or use other methods [02:12] that's amazing [02:13] along vaguely similar lines, for the chip datasheets I gave you, you could do a similar thing- on the first page or two, there's a set of bullet points about the chip's capabilities and features- that would make an excellent description [04:51] I gotta figure out git-annex very soon [04:51] * yipdw has an ssh key management mess [07:57] ssh key mgmt is always a mess [08:05] http://venturebeat.com/2014/06/11/now-comcast-is-working-on-a-youtube-clone-exec-confirms/ [08:05] just advance notice for something that will need to be archived when it inevitably dies [09:19] hum, i have a folder full of warc files, should i TAR them and then upload or upload them all as is? [09:59] ivan`: that's a new one, something being put on deathwatch before it's launched [10:01] "Comcast isn’t the only major company interested in launching a YouTube clone. While nothing has been confirmed, Yahoo is also reportedly working on a streaming video service to take on YouTube." [10:01] ... or perhaps not so new. [10:07] joepie91: your opinion, upload warcs in a tar or upload as is? [10:07] 1.4GB, couple of hundred [10:07] warcfiles that is [10:14] joepie91: Heh, yeah that's a thing now again [10:14] launching YT-clonez and "play services" [10:16] ersi: it's kinda funny, because Comcast basically says "it's like youtube, except not really about user-generated content, and we want to integrate it into set-top boxes, and..." [10:17] in other words: basically not /at all/ like youtube, but just a VOD service with content from non-giant studios [10:17] it's like you're listening to one of those "we need a developer to make this vision reality, we'll pay in company share" fools [10:18] midas: upload as is [10:18] that way wayback can ingest it [10:32] yeah, so it's basically a streaming service [10:32] ie a "play service" [11:16] joepie91: ok [12:19] woopwoop, uploaded all the homepages from XMSnet [12:20] it has pages like http://home.xmsnet.nl/caviastaljacqueline/ [12:24] And I actually find pages like that beautiful. It's like looking through a time tunnel. [12:26] What was the final size of the entire grab, out of curiosity? [12:35] JohnnyJac: which domain? [12:35] xmsnet? [12:35] 1.4GB in total :p [12:35] https://archive.org/details/home.xmsnet.nl [12:36] Not bad. [12:39] I am honestly just delving into using HTTracker for the first time. Done some stuff with wget in the past, though. Going to brush up on my skills because, in addition to running a dedicated Warrior rig, I'd like to contribute. [13:33] I wonder if i can see how much data my account uploaded to IA :p [13:35] For all time, you mean? [13:45] yeah [13:46] That would be interesting. [13:46] Wow. HTTracker is... addictive. [14:17] midas: you can use warcat or https://github.com/ArchiveTeam/megawarc [14:18] heh DFJustin, i mean, how much data i dumped on IA [14:18] yeah for your earlier question [14:19] you can see your total data uploaded with metamgr but I don't know if non-admins can access that [14:34] JohnnyJac: be aware that HTTracker does not output WARCs by default [14:35] what is this all about, no reply, mentions ' including 8Tbs of the shuttered geocities domain.' - http://www.reddit.com/r/DataHoarder/comments/27y8ux/standing_up_40tbs_of_data_for_fun_times/ [14:36] that's about the size of the archiveteam grab I think [14:38] er or not, maybe if it's uncompressed [14:40] I was using HTTracker for personal use to learn it better. I understand I would have to combine things into a WARC archive before uploading it. [14:41] well it's not just combining after the fact, you have to run the crawl through a warc proxy [14:42] Ohhhh! I learned something. I am nowhere NEAR ready to start contributing, but I figured I'll absorb some things just chilling in here. [14:47] DFJustin, the at get was around 700GB, that expands to 8TB :O [14:47] (?) [14:49] maybe they downloaded the rest of yahoo also ohhdemgir [14:49] like, all of yahoo. [14:49] XD [14:49] it happens! [14:49] jasons archiveteam talk mentioned a 9TB cage in relation to geocities .. hmm [14:49] (iirc) [14:52] Ha! I think that is the talk that led me here, actually. [14:54] :p [14:58] JohnnyJac: HTTrack doesn't preserve enough information to easily make a WARC, which you need to get something into Wayback [14:58] so look into that WARC proxy, or wpull, or archivebot (which uses wpull) [15:00] I will. I'll play around in spare time and get familiar with it. [15:03] or use wget with the warc flag :) [15:04] Hey, at least I have options. Hahaha! [15:10] Ah, so with Archivebot, I would need to be like, "Hey, guys, superawesomesite.com will be going down in a week. Can we archive it?" in #archivebot to have the command issued? [15:15] hmm sounds like it's less than 8tb https://github.com/despens/Geocities [15:15] so yeah I dunno what those other guys have [15:18] well if they have 8TB ... we (i) want it :3 [15:28] JohnnyJac: wget will also suffice, if you find yourself on a system that you cannot put other stuff on [15:28] just be sure to use the --warc-file flag [15:28] ivan`: is wpull production-ready now? [15:29] I was just curious if that was the mechanism to use ArchiveBot. I'd rather use wget anyway. [15:29] oh, was refering to the creating-a-WARC thing [15:29] as for using archivebot... regulars tend to get +o so they can issue commands to the bot [15:29] but you can always ask in there if you don't have +o [15:31] Burned an openSUSE 13 disc last night to get a box going for all of this. [15:31] Yeah, just exploring options. [15:31] whee! [15:31] :P [15:38] I am not a dunce when it comes to things, but I just have to orient myself to find out the procedures available to help out. [16:24] http://dc-js.github.io/dc.js/ is addicting [16:25] Dude... that is awesome! [16:26] it really is [16:27] So pretty... [16:35] ah yeah [16:35] I <3 d3 [16:36] I've used it for bioinformatics and telemetry work; it's a bit more complicated than e.g. highcharts but so much more powerful [16:57] "MediaRange is known for award-winning innovation in our market: With CD-R TRUE GOLD ARCHIVAL GRADE discs, we established a series of professional CD media which have a life span of 300 years and therefore comply with legal requirements regarding data storage for businesses. " [16:57] * joepie91 snorts [16:57] http://www.glassybaby.com/handmade-glass-votive-candle-holders-blog/wp-content/uploads/2013/06/I-love-gold-goldmember.jpg [16:57] lol [16:58] if you're not satisfied at the conclusion of 300 years we offer a full refund [16:59] hahaha [16:59] DFJustin: they do offer a "lifetime guarantee/warranty"... [17:35] waurantee [17:40] I can get a good look at a t-bone by sticking my head up a cow's ass, but I'll take my butcher's word for it [17:58] exmic: wuawuawuarantee [19:00] so, the ALDI is selling this: http://owely.com/41jBzUb [19:45] who the hell is medion [19:51] no idea [19:51] where is medion showing up? [19:59] so, the ALDI is selling this: http://owely.com/41jBzUb [20:55] DFJustin: medion is a budget brand in europe [23:56] after hearing "Turn Down for What" recently, I think that could be the unofficial ArchiveTeam theme