#archiveteam 2014-01-24,Fri

↑back Search

Time Nickname Message
00:00 🔗 Smiley lysobit: nope unless i win lotto.
00:00 🔗 BlueMax >__>
00:03 🔗 Smiley to -bs!
00:14 🔗 arkhive SketchCow: I am still going to send the stuff to you(floppies, manuals and such) i just haven't had time yet.
00:15 🔗 arkhive well.. haven't made time.
00:15 🔗 arkhive big difference. lol
01:40 🔗 RedType_ "Archive Team: Dammit Yahoo! And on the carpet too?!?"
01:53 🔗 yipdw RedType_: given the number of server problems these grabs are starting to cause, I'd suggest the small variation of "Archive Team: We Are Going To Fuck Up Your Shit"
02:55 🔗 xmc Archive Team: Smash & Grab & Xerox
03:42 🔗 BlueMax Archive Team: Sudo Save Your Shit
06:44 🔗 joepie91 Archive Team: Five Years Ago They Laughed
08:05 🔗 arkhive would be cool to rip these http://www.ebay.com/itm/The-video-encyclopedia-of-physics-demonstrations-1-index-25-book-w-Laser-Disc-/271127397864?pt=US_Texbook_Education&hash=item3f2073c5e8
11:16 🔗 SketchCow Great idea, or greatest idea?
11:18 🔗 SketchCow 5 years of Archive Team
11:24 🔗 midas no cake?
11:24 🔗 SketchCow Just got accepted - I am now running the Game Preservation SIG at GDC 2014.
11:25 🔗 SketchCow In other news, I just got a $995 pass to GDC 2014 for free
11:27 🔗 midas SketchCow: where should i send the cake, 300 Funston Avenue?
11:29 🔗 SketchCow Those people get enough cake
11:29 🔗 midas lol
11:29 🔗 midas it's going to be bloody hard to send a cake to everybody. hmm we need food replicators
11:32 🔗 Nemo_bis midas: they're called recipes
11:32 🔗 Nemo_bis yw
12:12 🔗 midas haha Nemo_bis ;-)
17:16 🔗 yipdw just revising this: http://archiveteam.org/index.php?title=Google_Video
17:16 🔗 yipdw er, revisiting
17:16 🔗 yipdw it's funny how big 18 TB is in the context of now
17:17 🔗 yipdw e.g. http://tracker.archiveteam.org/wretch/
17:18 🔗 Nemo_bis I've read that again too recently
17:19 🔗 Nemo_bis the news coverage was fun
17:40 🔗 Nemo_bis SketchCow: emijrp wants to write a paper on wikiteam; you once mentioned it would be nice to have a project like wikiteam but for forums, do you think that an export feature for forums is something important to advocate for?
17:42 🔗 Nemo_bis nowadays the biggest "forums" are Q&A sites and reddit-like things probably, some of those have an API (e.g. stackexchange)
18:22 🔗 SketchCow There are lots and lots and lots of focums that are neither Q&A and reddit
18:27 🔗 SadDM As a table-top gamer, I concur... most of my community's folk-knowledge is tied up in forums.
18:43 🔗 Jonimus Lots and lots of forums are still simple PHPBB and SMF boards with decades of content that can just disappear.
18:53 🔗 SketchCow Exactly!
18:58 🔗 Nemo_bis Sure
18:59 🔗 Nemo_bis I only meant, they have some pseudo-competitors with export functions, so maybe that could be an argument to convince them too
18:59 🔗 Nemo_bis though of course that won't help with the old forums nobody is upgrading
19:22 🔗 Jonimus Nemo_bis: most paid forum software provide tools to scraping their competitors HTML and building a new post database from that.
19:22 🔗 Jonimus I've actually done so to save an invisionfree forum.
19:22 🔗 Jonimus Its not as good as a DB dump but its better than nothing
19:23 🔗 Jonimus and of course the archive.org method of scrape all the things works fairly well unless there were sections of the site that are member only
19:23 🔗 RedType_ it's not even the sites going offline that you have to worry about
19:24 🔗 RedType_ all it takes is for an exploit to be released, and they're all fucked
19:24 🔗 DFJustin archive.org's standard crawl has pretty poor forum coverage in my experience
19:24 🔗 yipdw this is why we dump forums into archivebot :P
19:24 🔗 yipdw soon it will get better at that job
19:25 🔗 yipdw right now it's a gamble on whether you're going to run out of memory
19:30 🔗 Kenshin yipdw: 128GB RAM server ftw? :P
19:30 🔗 yipdw heh nah
19:30 🔗 yipdw the main problem with archivebot doing forums is that wget stores its URL graph in memory
19:31 🔗 yipdw chfoo's wpull can use an on-disk database, which should provide acceptable performance
19:31 🔗 yipdw and (I think) should give us constant memory usage per process
19:31 🔗 Kenshin so why worry about memory
19:31 🔗 yipdw because archivebot isn't using wpull yet
19:31 🔗 Kenshin ah ok, thus needing more memory
19:32 🔗 yipdw yeah
19:52 🔗 mietek WHAT FORSOOTH, PRITHEE TELL ME
19:52 🔗 mietek Requesting help archiving http://www.aux-penelope.com which is due to go dark on Feb 11, 2014
19:53 🔗 xmc mietek: yahoosucks
19:53 🔗 xmc penelope is going down?!?
19:53 🔗 mietek Check it: http://www.aux-penelope.com/aux_3.0.htm
19:55 🔗 mietek Aha! Scott Kanne responded to my email
19:55 🔗 DFJustin looks small, I sicced the archivebot on it
19:55 🔗 mietek > Thanks for the kind words - I have added a tarball of the entire site here:
19:55 🔗 mietek www.aux-penelope.com/aux-penelope.tar.gz
19:56 🔗 sep332 sweet
19:59 🔗 RedType_ someone should try and grab that domain
19:59 🔗 RedType_ it's over 10 years old, it's basically gonna get perma parked if scott lets it go
19:59 🔗 DFJustin https://twitter.com/ATArchiveBot/status/426805475355938816
20:00 🔗 DFJustin hmm true
20:00 🔗 RedType_ i can email him
20:00 🔗 RedType_ shit i'll even pay for the first year and transfer it to SketchCow or someone
20:01 🔗 mietek ++
20:04 🔗 RedType_ i know that's a dangerous (quickly expensive) road to go down, but even keeping it alive and point at archive.org for a year with a 301 would make a huge difference
20:05 🔗 mietek > I'm planning to keep the domain for now... I may pass it on at a later date though
20:05 🔗 RedType_ mietek: shit
20:05 🔗 RedType_ i just sent the email
20:05 🔗 RedType_ well at least he knows people are concerned
20:05 🔗 mietek Yep
20:05 🔗 RedType_ *donk*
20:07 🔗 RedType_ http://christtrekker.users.sourceforge.net/doc/aux/faq.html
20:18 🔗 Nemo_bis Jonimus: hah, competition, what a nice thing :)

irclogger-viewer