#archiveteam 2012-07-14,Sat

↑back Search

Time Nickname Message
00:10 🔗 Nemo_bis I've finished reading all the threads in [[Support]] from the beggginning of time \o/
00:10 🔗 Nemo_bis that was about https://translatewiki.net
00:12 🔗 Nemo_bis Coderjoe: are you sure about it? I'm almost certain that it once worked for me, saving me some Gbs of upload
00:56 🔗 chronomex greetings from HOPE
03:21 🔗 godane underscor: did you get these: http://www.demonoid.me/files/?uid=3783180&seeded=2
03:22 🔗 godane imaginefx 49-74 dvds
03:22 🔗 godane saidly i think its the contents of the iso and the iso's themself
03:22 🔗 godane but not sure
03:35 🔗 godane i'm planning on getting them now
05:36 🔗 Foxhack Um, hello?
05:37 🔗 Foxhack Is there anyone who can help me out? I'm the owner of a couple of sites that are listed in the Parodius Networking project.
05:59 🔗 Zod_ channel is a little slow, especially at this time
08:44 🔗 Schbirid anyone gave planetphillip.com a try? i will try to let it pass
16:33 🔗 godane uploaded: http://archive.org/details/cdrom-3d-world-118
16:33 🔗 godane uploaded: http://archive.org/details/cdrom-3d-world-115
17:40 🔗 godane uploaded: http://archive.org/details/cdrom-3d-world-124
17:43 🔗 ersi uploaded: http://archive.org/details/cdrom-3d-world-199
17:45 🔗 godane very funny
17:48 🔗 ersi :)
18:45 🔗 godane uploading this now: http://archive.org/details/cdrom-inside-mac-games-jan-feb-1995
18:45 🔗 godane its a hfs iso image
19:39 🔗 godane finally was uploaded: http://archive.org/details/cdrom-inside-mac-games-jan-feb-1995
21:07 🔗 Coderjoe we don't need the status updates, really
21:08 🔗 underscor #archiveteam-godane
21:08 🔗 Coderjoe Nemo_bis: well, I think for it to be able to resume, it would need to do a HEAD request to see what's already been uploaded, and the last time I poked, the s3api did not give the correct size in the HEAD response
21:09 🔗 Coderjoe (actually, I think it might have been nginx's fault, but I am not sure)
21:09 🔗 underscor s3 in general doesn't support resume like that, afaik
21:09 🔗 underscor Anyway, use multipart and send the missing pieces at the end
21:10 🔗 underscor https://gist.github.com/764224 this is what we use internally for a lot of stuff
21:10 🔗 Coderjoe well, you have to find out SOMEHOW what the current state is in order to resume.
21:10 🔗 Coderjoe particularly across multiple runs
21:11 🔗 underscor Well, yeah. In that case, you have to look and see which pieces are missing, and send those, and then send the multiput finished command
21:11 🔗 underscor it's not very clean
21:12 🔗 underscor anyway, if you use that python script, modify line 162 to:
21:12 🔗 underscor s3 = boto.connect_s3(key, secret, host='s3.us.archive.org', is_secure=False)
21:12 🔗 underscor and it will handle automatic retries and stuff
21:13 🔗 Nemo_bis Coderjoe: yes, dunno
21:13 🔗 Schbirid i am retrying planetphillip :(
21:28 🔗 Coderjoe meh
21:28 🔗 Coderjoe I was not aware that amazon had to make multipart stupidly more complext
21:28 🔗 Coderjoe -t
21:36 🔗 Coderjoe well, I suppose this makes sense for multipart
21:36 🔗 Coderjoe but for resuming, without parallel uploads, multipart is stupidly complex
21:43 🔗 underscor yeah
21:43 🔗 underscor :(
22:42 🔗 omf_ shit I was working on planetphillip too. Duplicate effort. I will stop until Schbirid gets back

irclogger-viewer