[01:41] probably noted already but arstechnica.com/gadgets/2012/06/minitel-frances-precursor-to-the-web-to-go-dark-on-june-30/ [02:45] huh I remember learning about that in french class [04:05] godane: http://archive.org/details/cdrom-3d-world-140?reCache=1 [04:05] got the other one too [04:05] I got collection moving privs today [04:05] hooray! [04:07] You're all grown up [04:08] hahaha [04:08] you can move these: cdrom-3d-world-120, cdrom-3d-world-121, cdrom-3d-world-116 [04:08] I got the warning from my boss, "Don't accidentally fuck something cause it makes me look bad!" [04:08] hahaha [04:09] and: cdrom-3d-world-125 [04:10] godane: done [04:11] thanks [04:11] our cd-roms will blot out the sun [04:13] Then we'll archive in the shade! [04:14] I hadn't seen those cd stack photos from jason's presentation before though, holy piss [04:14] lol [04:20] lol [04:53] iirc, jason said "just upload them, and I'll move them when you're done" [05:15] uploading dltv episode 73 [05:15] my brother is watch late late show [05:15] so i can upload quick stuff now [05:15] :-D [05:16] sometimes when I wake up, it feels like I woke up from death. ugh. I seem to have forgotten what quality sleep is. [05:16] lets see if these isos are ready to upload [05:19] Does anyone use FXP for uploading files? I was thinking of giving it a shot [05:19] matrix is back [05:19] watch episode 73 [05:22] Coderjoe: yeah, but might as well save him effort [05:24] I'm in LA for the next 3 days. [05:26] Uploading a previously missing 6.9gb of Geocities [05:26] there was 6.9gb missing from geocities [05:27] Where'd you find it? [05:30] in the sofa cushions [05:30] it fell out of someone's pocket [05:30] Yahoo HQ dumpster [05:31] Was on a machine from a volunteer server that went down. [05:31] The guy got it up, and it had this data, and some google videos too. [05:32] god I suddenly made myself feel old. I remember when 6.9GB was in incomprehensible amount of data. Now you can carry it around on something the size of your pinky fingernail and have it fall out of your pocket and loose it [05:32] Pull out your wallet, lose Geocities [05:32] (holy shit! I have a 15MB hard drive! this is awesome! I have no idea how I can fill this up!) [05:39] Coderjoe: I always carry 8GB with me on my keychain [05:41] http://en.wikipedia.org/wiki/File:IBM_card_storage.NARA.jpg is my favourite comparison image [05:42] I now have a microsd card that holds 8 times as much as that warehouse [05:43] And 8 million times easier to get data from [05:46] Hah, I love how an offhand calculation I did in irc made it to the description of that image [05:46] We truly live in the future [05:47] woop woop woop off-topic siren [05:48] ersi: yeah? these are usually in my pocket: https://dl.dropbox.com/u/71339106/in_pocket.jpg [05:49] that's right, I carry 76.5GB around in my pocket, not counting my cell phone [05:51] Just blew more mobileme in. [05:55] before there was video and mp3 it may have been hard to think you could fill 15mb hard drive [05:56] most people only could do word processing or use ms paint to really file the 15mb drvie [05:56] *drive [05:57] actually, the first hard drive I had was a 10MB one that was factory-mounted onto the ISA controller card. [05:57] Awwwyeah [05:58] which was put into the family AT&T PC6300 (an 8086 with 640KB of ram and two 360KB floppy drives) [06:00] I had an IBM XT with a 20mb ISA slot hard drive [06:00] and it still boots and works today [06:00] now I do feel old :( [06:00] -bs [06:01] With the end of the mobileme run, we'll have no archiveteam download projects going. [06:01] I mean, I'm glad, but are we missing anything? [06:01] SketchCow, You had a question for me earlier on the linux isos I am backing up [06:01] I think it was asking if you are scanning the cds and the labels [06:01] SketchCow: Fileplanet's finishing up, esp. since they're cracking down on us [06:02] But nothing that needs concentrated effort [06:02] Oh, fileplanet. [06:02] So are we going to miss a large percentage? [06:02] I don't think so [06:03] Searching for off-the-list stuff comes up with a few stragglers, but nothing big [06:03] shaqfu, how many were missing from that chuck of 5k urls you fixed? [06:03] We probably got 95%+ [06:03] omf_: Only 500 URLs, and not many [06:04] that is good [06:04] It's kinda funny; that's the second time they tried stopping us from wget'ing, and the second time it was fixed in 15 minutes [06:04] add episodes 73 and 74 of dl.tv [06:04] SketchCow, for the disks I have I plan to scan the disk cover, cd case cover and back and the insert if they have it [06:08] does anyone have an estimate of the total size of fileplanet? [06:08] omf_: Thanks [06:09] I assume you also want all the isos I have that don't have that stuff as well. I did some searching and I can find that stuff around to add later [06:09] omf_: You'd have to sum the list on the wiki [06:09] I will do that shaqfu [06:10] 5TB sticks out in my mind, but dunno if that's accurate [06:31] OK, I appear to have found hundreds of gigabytes, maybe terabytes, of Yahoo Video. [06:31] So that's going up shortly. [06:40] SketchCow, I created an a.o account. I assume I cannot create a collection to put these disks in? [06:41] correct. collection creation needs administrative action [06:41] will you guys have problems with me uploading x-play? [06:42] Look, just keep uploading. [06:42] i have maybe full 2011 year of shows [06:42] ok [06:42] I wish you would upload individual shows as individual items. [06:42] I may break apart some of your items in the future. [06:42] i am with isos and dl.tv [06:42] and crankygeeks [06:43] the screen savers i did that why so i didn't go insane uploading one episode at a time [06:43] Famicoman as well [06:43] godane: you can use s3 to script it :) [06:43] i don't know how [06:44] also stuff like dl.tv has a lot of info i want to add [06:44] like links to stories [06:45] yeah, header length limits kinda makes it painful for long descriptions [06:45] metadata.xml ! [06:46] also when i uploaded stuff like gbtv put up like 5 episodes so i can walk away from it for like 5 hours [06:46] http://archive.org/details/GBTV_09_12__16_2011 [06:46] Do you use unix? [06:46] i did this cause other glenn beck stuff is in month packs [06:47] yes [06:47] custom linux [06:47] You should really be using s3. [06:47] i have a python script I used to upload the stage6 content using the s3api [06:48] And I use bash. [06:48] I should do a round of making my scripts more obvious. [06:48] (which was mainly because I was pulling in info from mysql for the metadata. it just called out to curl to do the work) [06:48] with dl.tv its special [06:49] like i want time length and all the info and links with keyboards of who is the host and guests in that episode [06:49] and that can change a lot [06:52] so you can put that info in a text file alongside the show media file and have the script read the info from the file [06:53] where is the s3 code? [06:53] i don't even know where it is [06:54] thats my main reason for not using it [06:54] http://archive.org/help/abouts3.txt [06:58] I have scripts, you should use them [06:58] I'll do a round of revision. [07:02] i found this: https://github.com/kngenie/ias3upload [07:05] godane, that script almost made me blind [07:08] in just 2 min [07:08] i agree [07:09] is it the same as https://wiki.archive.org/twiki/bin/view/Main/IAS3BulkUploader ? [07:10] So, how many times do I have to say that I have written scripts I use internally to upload tens of thousands of items? [07:10] I mean, I guess I can make the atlogbot do it [07:10] SketchCow, it's not a bad idea [07:11] this guy has some such IRC bots: https://github.com/mzmcbride [07:11] SketchCow, where can I get your scripts from [07:12] i can't download from wiki.archive.org [07:15] I need to do a round of revision. [07:35] i'm looking at the ias3upload scripts [07:36] i'm starting to not like s3 cause its going to cause a headache for me [07:36] Goddamnit. [07:36] Why are you so productive, and still a dope. [07:37] I've been up for 18 hours now, so I'll deal with things tomorrow. [07:37] sorry [07:37] i just thing i will screw things up [07:37] *think [07:37] nah, it's very easy [07:38] plus, you can always start over [07:38] i still don't see how to do it [07:38] very little room for mistakes [07:38] you only have to edit the csv with essential metadata and run the script [07:38] i do ids like dltv_$num_episode [07:39] also will i have to add a new id for each one [07:41] that is not complicated [07:41] also you can blame the autium i have [07:42] i sometimes understand stuff in very different way [07:46] even if you are a dope sometimes, you're doing things. doing things matter more [07:46] like you grabbing all those shows and stuff, it's all ok [07:46] * ersi pats godane [07:47] i'm also getting all 3d world iso i can find [07:47] yeah production matters more than anything [07:47] hell I am about to do my first upload and I am floundering around a little bit [07:47] thats out of normal for me as far as archiving [07:48] my first upload was a floss weekly episode 80 video [07:49] http://archive.org/details/FlossWeekly80_ScummVM_live [07:49] the video was the live stream cap [07:49] not the episode video [07:50] funny [07:51] google search of floss weekly 80 video cames up with my archive.org one [08:12] So what is the next AT project? Want me to put up a website and threaten to take it down? :P [08:13] with TBs of 000000000000 [08:14] We could work on getting archivial methods for other, more popular websites if their d-days ever come [08:14] we could work on a way of discovering sites that are closing so we can look into if they need backups [08:15] isn't one of the big problems not finding before a site shuts down? [08:16] correct [08:16] i some how don't know how to use .csv [08:20] BlueMax, any sites in specific? [08:20] omf_, I was thinking social aggregation sites, like Digg or Reddit [08:21] I'm working on my own Wayback Machine sort of [08:22] BlueMax: Sounds like something for URLteam imo (I think aggregation sites like those are ultimately urlshorteners with possible extra metadata(comments+comment scores)) [08:22] ersi I see your point [08:23] I already backup sections of reddit [08:23] some subreddits I have all the way to their beginning [08:23] neat [08:23] I am trying to get some back data from them as well [08:23] I use it as a source to find things for nlp. This is what drove me to AT because of the fanfiction.net proces [08:23] where else can I find literature to work with [08:25] what about flickr? [08:25] when yahoo closed 'yahoo photos' there was 250 million accounts [08:25] got #flickrfckr + project [08:25] it's a bit idle, but there's scripts and stuff AFAIK [08:27] Myself as well as others already download the wikia backups, stackoverflow backups, wikipedia backups, etc... [08:28] what about fotopedia.com? [08:29] if any project interests you, feel free to hax on it :) I have never heard of fotopedia.com myself [08:29] if anyone wants to snag http://www.sfseriesandmovies.com I would really appreciate it. My internet is only 8megabits down at the best and I am trying to get this linux iso archive built [08:29] they have some great production photos [08:37] i've got some walnut creek slackware cds out in storage. only the cds, though. none of the printed docs. [08:37] Coderjoe, awesome [08:38] slack was my first distro [08:38] same here [08:38] I got an old 4 cd set with slack, sls, debian and others [08:40] i downloaded slack disks off a local bbs whose sysop was kind enough to download them at school and put them on a filebase on his bbs [08:41] however, i don't think it was the first unixy system i ever used [08:42] that would go to a sysvr4 system i dialed into, which only had uucp email at that point. [08:42] woopwoopwoop [08:56] http://www.ballyalley.com/ [08:58] Schbirid, cool site. should we back it up [08:58] no idea, just randomly got on it [09:01] i can get s3 working in bash [09:01] :-D [09:02] yes, isnt it awesome? [09:02] how do you tell curl to give output on how much its uploading? [09:02] like speed and stuff [09:03] curl '-#' [09:05] i'm only doing one for now [09:05] but this may work better with my isos anyway [09:06] dl.tv needs a lot more info [09:10] or -o output.txt [09:12] there will be no output unless there is an error, but curl defaults to outputting the retrieved data on stdout. adding the -o will cause it to write that to a file and output the upload/download stats instead [09:15] this item is not found: http://archive.org/details/dltv_076_episode [09:15] thought it would have checked it in [09:15] is the upload finished? [09:15] yes [09:15] though the item should have been created not too long after starting [09:16] did curl report an error? [09:16] no [09:16] http://archive.org/catalog.php?history=1&identifier=dltv_076_episode [09:16] oh, be patient [09:17] there appears to be a bit of a backlog going on [09:21] http://arstechnica.com/gadgets/2012/06/minitel-frances-precursor-to-the-web-to-go-dark-on-june-30/ ! [09:26] hi there [09:27] hello commander [09:29] Schbirid: btw, that link was already mentioned: #archiveteam.log:2012-06-28 21:41:03EDT by BlueMax [09:29] ah ok [09:29] is there a reason why the archiveteam warrior is not linked from the wiki's software page? [09:30] and sadly, I don't think there is anything that we can do about archiving it [09:31] C-Keen: the wiki is a bit neglected [09:32] the wiki was moved awhile back and some boxes of connective parts is still elsewhere [09:32] :D [09:32] what did i do wrong now: http://www.us.archive.org/log_show.php?task_id=110806334 [09:33] please free this up so i just do it the ftp way [09:34] Warning: fileperms(): stat failed for /13/items/dltv_076_episode in /usr/local/petabox/www/common/Util.inc on line 651 [09:34] Unknown err: file_get_contents(http://archive.org/services/find_file.php?file=dltv_076_episode&loconly=1): failed to open stream: HTTP request failed! [/usr/local/petabox/www/common/ItemTools.inc:28] [09:35] ok so the item id is in use [09:35] I can't do anything to fix it. I really have any admin powers at IA. that error makes me think there might have been a hiccup in the system. There appear to be some nodes that are marked read-only, and I don't know if that affected this [09:36] this is why i going back to ftp [09:37] can you PM me the curl command you used (replacing your s3 key info) [09:37] there doesn't appear to be an item with that identifier, other than your attempted upload [09:39] i tryed creating dltv_076_episode [09:39] it said: Desired identifier "dltv_076_episode" not available. Please try another identifier. [09:39] so it must have been used [09:40] you created it [09:41] once you've done the amz-create-bucket request, you can't create it again. you can upload files to an existing bucket, or (using another header flag) tell it to discard the old metadata so you can respecify it [09:44] that command looks correct, for the initial creation of an item anyway, as long as all of your variables are filled in [09:44] but you're not going to be able to do anything with that identifier until an admin fixes it [09:49] ok [09:51] moving on to episode 77 for now [09:51] will use ftp from here on out [09:52] you said its simple and thing it acts like that i just can't trust s3 now [09:53] and what will you do when proftpd craps itself? [09:53] cry [09:54] i'm just hoping its just my stupid script at this point [09:54] seriosly, SketchCow has uploaded tens of thousands of items. I've uploaded around 5000 items. [09:55] the seesaw-s3 script that was available to high-throughput memac contributors has uploaded a lot of items [09:55] i just think there is some vodo code that i didn't add to my script [09:59] it looks like the dltv_076_episode folder was never made [13:25] godane: Fixed (again): http://archive.org/details/dltv_076_episode [13:25] Why am I fixing other people's stuff lol... [13:30] Are these sites http://archiveteam.org/index.php?title=Fire_Drill also proactively backed up or is this just a monitor list? [13:32] I know a few of them have been backed up [15:25] Is there like an 'overlord' for the wiki or do we just fix as needed? [15:26] omf_, it's a wiki so everyone can fix things ;-) [15:33] * SketchCow is now pasting in BBC monograph descriptions. [15:46] "Unfortunately, the time has come to say goodbye to Yahoo! Groups Labs Applications (beta) and Groups Chat. We thank all the users that were a part [15:47] of this program. We intend to use this learning to enhance Yahoo! Groups further with new features. [15:47] On July 04, 2012, we will shut down the Yahoo! Groups Labs Applications (beta) and Groups Chat. We request you to backup any data that you might have [15:47] created using the applications." [15:48] what, 5 days of notice!? [15:49] Do we have any URLs for these things? [15:53] It's probably mostly private data (only for members of groups) but I don't know [15:54] Take a look around, if you could. [15:54] Sure [15:54] I agree, it's likely not public, but we can see. [15:56] Applications are stuff like address book, birthdays, favorites etc. that can be enabled by the group owner [15:56] Chat, I can't find in the UI [16:08] ooh I think I found something. It is a hosting company's userpages which is like a mini-geocities. It has 50 different sites that all appear to be from the 1990s [16:08] I am totally geeking out [16:10] time to download all of it [16:14] netlabels are so volatile :( [16:15] the wiki just cried out: [16:15] Fatal error: Call to undefined method Article::getSection() in /home/archivet/public_html/extensions/recaptcha/ConfirmEdit.php on line 620 [16:17] yeah, you cant edit sections, only the whole page (or you will get that bug) [16:17] aah [16:18] Bug I need to fix. [16:18] The new wiki is very secure on the backend (we've had no spam hacks, you've seen), but it makes it somewhat pain in the ass to fix things. [16:20] secure is good [16:21] I've seen a lot of spam users on http://www.archiveteam.org/index.php?title=Special:RecentChanges but I don't think those are spam hacks :) [16:23] No, that's because they're defeating captcha. [16:24] What captcha system is in use? [16:42] hmmm how many new users per day say, does the wiki get? [17:09] Just got off the phone with a CNET reporter, we'll make an appearance in the MobileMe shutdown story. [17:09] http://archive.org/details/bbc-rd-reports-1996-13 [17:10] I have 1000 of these to describe. [17:12] hahaha 1,337 items [17:50] SketchCow: We should have a game-ified "describe this item" interface [17:50] with trusted users who can "verify" descriptions too [17:50] (like, get 20 descs approved, gain the ability to approve others' descriptions) [17:50] then various IA teams could just dump things into the "pool" [17:52] ugh [17:52] this is why i going back to ftp [17:52] FTP would have broken WORSE in that situation [17:52] It happened because an exhaust fan over one of the server racks got shut off [17:52] (breaker tripped) [17:53] so that rack got excessively hot, and was triggered for auto OFFLINE and poweroff by nagios [17:53] And it just caused a few locator requests to fail because catalogd was in a state of limbo [18:21] that's a good idea [18:23] personally I think all the IA item metadata should work like say musicbrainz where anybody can edit and empowered users approve but that's probably a harder sell [18:24] because the guy who is most knowledgeable about item x and cares the most about fixing it is probably gonna be some guy who finds the page on google and not one of us [18:34] Seems fruitful, although I dunno of any large-scale crowd description efforts [18:34] usually orgs use the public as a free MTurk [18:35] But it's worth trying; dunno of any other sane way to describe IA [18:36] As long as the whole thing has a viewable history and rollback (just in case) I don't see any reason why that isn't a fantastic idea [18:48] How much are we looking to raise off Kickstarter for drives? [18:59] http://ascii.textfiles.com/archives/3642 [19:00] Personally? I'd like to see us get to $100k [19:01] How much space does that get? [19:02] SketchCow: a tour of IA would be a neat perk, I think, though I don't know how that works out with travel costs (and whether IA gives tours anyway) [19:07] for a 100k$ donation I think they could arrange a tour [19:07] underscor: Speaking of archive-by-fans, Grateful Dead went up today [19:08] Speaking of archiving, I met some cousin of my best friend yesterday and she's /teaching/ digital archivism at whatever her college is [19:08] Aranje: Which city? [19:08] I have no fucking idea [19:08] I told her to find archiveteam and drop in here sometime [19:10] http://www.archiveteam.org/index.php?title=Kickstarter_Prizes [19:11] The basic prize should be something like "get your name on a special list of contributors to IA for all eternity" [19:27] IA has swag? where can I purchase all of it [19:37] hmmmm makerbotted things [19:37] speaking as someone who's operated a makerbot, don't commit to making too many items [19:38] High rate of failure? [19:39] slow, pain in ass. [19:42] it seems a little more realistic than the device itself because it's so expensive [19:42] unless you raffled it or something [19:43] Not sure if raffle tickets make good KS prizes [19:43] Unless it's part of something else [19:44] selling raffle tickets for anything other than a registered non-profit entity is a felony in some states [19:44] just fyi [19:46] We could do what unis like to do and sell naming rights :P [19:47] "URL Shortener Archiving Subcommittee Brought To You By Some Internet Guy"? [19:47] "The Joe Smith Collection" [19:48] For the low low price of $100k [19:51] perhaps obvious, but I'd suggest curated collections of interesting archives. All of textfiles.com (well SketchCow used that as a KS reward before), the curated/cleaned up geocities archive from the 1tb blog, etc [19:52] data that's on archive.org, but you add value by putting the collection together, and putting it on media [19:56] hmmmm, yeah [19:58] archiveteam warrior on a bootable cd would also perhaps work as a $25 or so reward [20:00] and also on a bootable udb key [20:01] good thinking [20:02] 'd have to modify it to use tmpfs for storage, but that shouldn't be hard [20:02] you can get a 4 gb Really Nice usb key for around $5 in bulk [20:03] with printing [20:04] oh and put an autorun virtualbox on it too :) [20:04] looks more like malicious software :p [20:05] "oh, what a nice collection of websites here! Hey, wait, what happened to my internet connection?" [20:23] Aranje: We don't have a great rollback system atm [20:24] and there's probably not enough manpower to do code it up [20:24] yipdw: We give tours every friday [20:24] Free lunch and tours and meet and greet [20:24] as long as you're willing to sign the guestbook [20:24] :D [20:25] where is more info btw? [20:30] about? [20:35] about visiting, I suppose [20:39] I would gladly trade a HDD against some Internet Archive or Archiveteam swag like a T-shirt [20:45] anyone have a dropbox referral link they want me to use? not paying for an account just yet, fwiw. [20:54] http://db.tt/4sH91ZXl [20:54] only 10 minutes late [20:54] Famicoman: say his name at least [20:55] * Famicoman summons chronomex [20:55] Famicoman: too late, someone else beat you to it in another channel [20:58] we were semiseriously thinking about a prize being an old HD signed by the board/brewster [20:58] Because we have stacks of 250-750GB drives that just sit and do nothing [20:58] Dead drives or just old drives? [21:01] just old [21:01] Put them on Ebay? [21:01] left over from 500/750GB -> 1TB -> 2TB -> 3TB migrations [21:01] why would any sane person buy used hard drives [21:01] ^ [21:02] especially ones abused by us [21:02] it's like buying expired food [21:02] 24/7 high bandwidth access [21:02] No idea, but there are actually many people selling/buying old hdds [21:02] all the buyers are idiots [21:02] all the sellers have the right idea [21:02] there are also people who buy and sell used schoolgirl socks on ebay. doesn't mean the buyers are sane [21:03] http://www.ebay.com/itm/Well-Worn-Used-Japanese-School-Girl-LOOSE-SLOUCH-SOCKS-from-JAPAN-/170865965537?pt=US_Womens_Socks&hash=item27c867b9e1#ht_8775wt_1398 [21:03] Why the hell would you pay $50 for a pair of socks?!?!?!??!?! [21:03] Actually it's not just Japanese schoolgirls, but all kinds of socks [21:03] You have to sell them as "clean, washed before sent" but in reality the dirtier the better [21:04] soultcer here is a seasoned buyer of ... seasoned socks [21:04] ...hahahahahahaa [21:05] I guess there are people with sock fetishes? [21:05] yes [21:05] woop woop woop off-topic siren [21:05] I knew it was coming [21:05] my wwwots-dar was going off [21:06] ... [21:07] :D [21:21] some fix my item [21:21] http://archive.org/details/dltv_076_episode [21:21] *someone [21:21] *fixed [21:21] :-D [21:22] someone named slaxemulator, it seems [21:22] http://archive.org/catalog.php?history=1&identifier=dltv_076_episode [21:37] i think i'm uploading dltv 079 [21:37] trying s3 again [21:37] but its doing some thing different [21:37] is it normal for there to be task when uploading? [21:37] http://archive.org/catalog.php?history=1&identifier=dltv_079_episode [21:38] i added '-#' [21:38] but getting no output for upload speed [21:46] ah, you need to direct the normal output to somewhere else [21:46] * chronomex digging up example [21:47] add also: -o /dev/null [21:50] how do i add runtime info? [21:53] ? [21:53] I don't think curl has a way to output estimated time [21:54] maybe --progress-bar [21:55] or not [21:59] so can i just add stuff to shareware cds now with s3? [22:00] or do i need specail admin for that? [22:03] you need to be collection admin [22:03] but it's not important [22:04] yes it does esteem time, but use > /dev/null or whatever [22:05] (otherwise it's most frustrating, days of upload without knowing how many GBs are left :) ) [22:08] The MobileMe article with us making an appearance: http://news.cnet.com/8301-13579_3-57463921-37/mobileme-a-rare-apple-screwup-finally-bites-the-dust/?tag=mncol;topStories [22:15] godane: Upload normally, one of us can frob it [22:16] (me|jason|other IA folx) [22:20] I've been playing with a new version of the warrior VM. So far it seems to be 100MB larger than the previous one (350MB instead of 250MB). I'm not sure if that's a problem. Preseed configuration files are here, https://github.com/ArchiveTeam/warrior-preseed, if anyone wants to see if it can be made smaller. [22:20] I noticed this in the keynote too, tripod hasn't actually been shut down [22:21] Yeah, I keep doing that. [22:21] alard: What's causing the larger size? [22:21] lots of manjuice included [22:21] >_> <_< [22:22] i mean, Archiveteam-juice [22:22] SketchCow: Not sure. It may be because it includes a few extra packages, more of Python, for instance. It may also be because I optimized the previous one better. [22:23] It's surprisingly hard to tell the debian installer that it shouldn't install certain packages. You have to remove them afterwards, it seems. [22:23] have you run one of those tools that zeroes out unused sectors and re-compacts expandable vm images [22:24] I did that on the first one, manually, and have included it in the preseed file: https://github.com/ArchiveTeam/warrior-preseed/blob/master/warrior-preseed.cfg#L758-760 [22:24] it's fairly simple: in the vm: dd if=/dev/zero of=bigfile ; rm bigfile outside the vm, use dd to sparsify it [22:24] I didn't know there was things to do it on a vm image specifically though [22:25] What are the new features/changes between old warrior and new warrior? [22:25] And then there is the VBoxManage --compact option: https://github.com/ArchiveTeam/warrior-preseed/blob/master/pack-vm.sh#L6-7 [22:25] well VMs use image formats with headers as to which sectors actually have stuff in them, so you need a tool that groks that format if you want to distribute the image in that sparse form [22:26] The new warrior (isn't ready yet) comes with Lua and Python. The big change is that it will have a web interface. [22:26] but anyway the answer seems to be "yes" :) [22:26] https://github.com/ArchiveTeam/seesaw-kit [22:28] The web interface will contact the warrior HQ, lets you pick a project to work on, shows stats and configuration options. I want to add a 'let ArchiveTeam choose a project' option, so you can leave your warrior running for a longer time. [22:28] All sounds good. [22:29] Ironically, we're out of projects tomorrow. [22:30] I guess a preemptive backup of Flickr might be fun [22:33] We can prepare for the next project. Get the shared tracker set up somewhere, for instance. [22:33] "They probably will make it available for something like $10/GB and apple gets 33% of the cut. I am sure they didn't do this out of the kindness of their hearts. Unless they are in it for the potential homemade porn." [22:34] oh macrumors users [22:34] The Macrumors post! [22:34] I love that guy! [22:34] yipdw: Is that in reference to memac? [22:34] shaqfu: yes [22:34] Hahaha [22:34] I dunno about you guys but I'm sure as hell in it for the homemade porn [22:34] I've archived that thread [22:35] DFJustin: same [22:35] memac-on-geocities.avi [22:36] I like how all of these comments on the macrumors threads are like "BUT I DIDN'T MEAN FOR THIS TO HAPPEN" [22:37] re: Public folder [22:38] SketchCow: there's some (admittedly non-tracker) last-minute stuff going on in #fireplanet [22:39] nothing to panic about imo [22:39] Right, we have others, too, but I just was wondering if we had any other focused tasks. [22:39] http://www.zdnet.com/blog/apple/lost-in-the-shuffle-mobileme-galleries/13109 [22:47] 15:33:29 <@alard> We can prepare for the next project. Get the shared tracker set up somewhere, for instance. [22:47] alard: I'll buy a linode and let you play with it, how's that sound [22:48] man, I love the archiveteam-warrior vm [22:50] me too :-) it's awesome [22:51] SketchCow: I'm not a dope cause i can use s3 now [22:52] still trying figure out how to get download speeds out put [22:53] chronomex: That might be fun, though the tracker could also piggyback somewhere. It doesn't use a lot of resources. (It just eats a lot of memory, as you've seen. :) [22:53] is 512M ram enough? [22:53] Certainly. [22:54] ok [22:55] tokyo, london, new jersey, georgia, texas, california [22:55] http://bits.blogs.nytimes.com/2012/06/29/the-well-a-pioneering-online-community-is-for-sale-again/ [22:55] Could someone look at that? [22:55] alard: have a preference? [22:55] No, I don't. [22:56] (Perhaps tokyo is far off from most people.) [22:56] London. [22:56] Do London. [22:56] ok [22:56] why, close to .eu? [22:56] Far from .us without being too far [22:56] ok [22:57] And our amazing devs are mostly eu right now [22:57] distro? [22:57] Debian? [22:57] k [22:58] hostname? [22:59] SketchCow? (Not as the hostname.) [22:59] hahaha [22:59] panopticon.archiveteam.org [23:00] too late, I named it 'shilling' [23:00] I guess that gives you a natural upgrade path [23:01] I name my hosts after currencies [23:01] SketchCow: where is your code for s3? [23:01] "dollar is being stretched too thin" [23:04] Hm, is it still possible to register for The Well? [23:04] If it's at-risk, we could get an account, share it, and start archiving [23:05] Oh, only $10 to join [23:05] we should do a kickstarter to buy it [23:28] i'm almost done with my 2nd dvd dl of dl.tv [23:28] uploading episode 84 [23:31] http://www.bbc.co.uk/news/magazine-18610692 Minitel is shutting down. [23:32] Is there anything we can do about it? [23:33] I'm pretty sure the answer is "no, unless we've got someone in france" [23:34] A shame - we have the resources on hand for a four-alarm save right now [23:35] what is a 5-alarm save then? [23:35] break in and clone hard drives? [23:35] heh [23:36] dumpster dive? [23:36] if there not just fully destoryed [23:36] Gain access to the machines and do everything to keep them on [23:36] While someone else chains themselves to the servers [23:37] using rabbit ears wifi to keep the internet connected [23:38] OK, I think fanfiction is now, I think, pretty much compiled.