[00:01] Hostgator just suspended me. :) [00:01] They want account verification. [00:01] I assume this is because I bought a 3 year business account. [00:01] epic [00:08] jeez, that was fast. [00:31] I'm refunding hostgator and switching to webfaction. [00:33] <3 [00:34] SketchCow: let me know if you have any questions w/ the webfaction setup [00:34] SketchCow: the app/domain mapping is a bit unique to them [00:36] I want to run a mediawiki on a domain. [00:36] It's not that hard [00:36] If it IS hard, something's wrong [00:39] He;;p [00:39] Hello* [00:40] Hi [00:40] Is there ssl here? I cant seem to connect to port 6697 [00:40] Oh, it's connecting now, sorry [00:41] Thank you for signing up with us. [00:41] I just bought a 5 year account at webfaction. [00:42] Is that the dealbreaker problem you were talking about on twitter? [00:43] The dealbreaker problem was hostgator suspending my account while I was working to move files to it. [00:43] And then demanding ID, and then taking an hour or two. [00:43] Sorry, so fired [00:44] Either demand my ID as a final step, or don't demand it. [00:44] SketchCow: i've heard good things about rackspace [00:44] Put it up, make it work, THEN jam it down, ESPECIALLY when I just gave you $350 bucks, sorry, you're shit and I'm done. [00:44] If webfaction shits the bed, rackspace is next. [00:44] have you tried namecheap? [00:45] I use them for domains, not hosting. [00:45] Ah [00:46] SketchCow: they actually have a one-click install for mediawiki [00:46] v1.8.2 [00:47] Good. [00:48] SketchCow: https://panel.webfaction.com/app/create?script_url=https%3A%2F%2Fwiki.webfaction.com%2Fwiki%2FMediaWiki-1.8.2Install%3Fformat%3Dtxt [01:18] Wow, that's one fucked up interface. [01:18] I wish you had told me it would be something by geeks to provide to nerds to make them all feel like sexless robots with infinite processing space and time [01:19] I want to hear the AWESOME metric this approach fulfills [01:19] I want it so you go to archiveteam.org and it shows the wiki. Looks like this thing tries to shove you into webapps. [01:19] SketchCow: i told you to ask :) [01:20] SketchCow: create mediawiki app, create archiveteam website, add architeteam domain to said website [01:20] and you'd golden [01:20] *you're [01:21] I like that. [01:21] 1. Turn lead to gold [01:21] haha [01:21] 2. Forge gold into statue [01:21] 3. Tell golem to emulate gold statue [01:21] 4. Instant sandwich! [01:21] I just want a sandwich [01:21] it's setup so you can have, say, wordpress running under a certian path on a domain [01:22] like mediawiki for /, wordpress for /blog/ [01:22] Yeah, it's on crack. [01:22] Anyway, you're no help [01:22] I obviously have to go through some sort of cult transition to think in this retarded geek environment [01:22] it takes 3 seconds to setup [01:22] So dropping spoilers that mediawiki is a sled isn't going to help [01:22] Yes, it takes me 3 seconds to find out the processing of a machine too [01:22] Unless I'm dropped into Unix like a foundling [01:23] Allow me some time to go read this crazy moonbat environment up [01:24] SketchCow: http://docs.webfaction.com/user-guide/websites.html [01:24] Too late, I'm working over here [01:28] gotta be fast to keep up with the cow [01:29] Error: Traceback (most recent call last): [01:29] File "/var/tmp/api_system.28582", line 96, in ? [01:29] func(account, app_name, autostart, extra_info, password, server, session_id, username) [01:29] File "/var/tmp/api_system.28582", line 58, in create [01:29] server.create_db(session_id, db_name, 'mysql', password) [01:29] File "/usr/local/lib/python2.4/xmlrpclib.py", line 1096, in __call__ [01:29] return self.__send(self.__name, args) [01:29] File "/usr/local/lib/python2.4/xmlrpclib.py", line 1383, in __request [01:29] verbose=self.__verbose [01:29] File "/usr/local/lib/python2.4/xmlrpclib.py", line 1147, in request [01:30] return self._parse_response(h.getfile(), sock) [01:30] File "/usr/local/lib/python2.4/xmlrpclib.py", line 1286, in _parse_response [01:30] return u.close() [01:30] File "/usr/local/lib/python2.4/xmlrpclib.py", line 744, in close [01:30] raise Fault(**self._stack[0]) [01:30] xmlrpclib.Fault: [01:30] Yeah, there's some aweseom [01:30] Well, thanks for dropping me into a true level of hell [01:30] It's like driving a car through a playground of babies and steering with PERL [01:30] I mean, a serious, level of hell [01:30] 16 characters or less. [01:30] Let's do surgery with algrebra. [01:30] welcome to 1983. [01:31] Shut up, once we used 17 characters in a filename and a man DIED. [01:31] He DIED. [01:31] was it CP/M? [01:31] cause serves him right if it was. [01:32] I heard he was trying to use sed and awk into his perl compost because he was cooking a muffin [01:32] Luckily his C++ preprocessor was fibbling the foorogotz and we didn't blow up a nearby church [01:32] This is all in the script [01:33] is the mediawiki in the debian/ubuntu repos not an option for some reason? [01:33] for all i know it's really old or no one uses it or it isn't there [01:33] Your account will be canceled in the next 24 hours (usually much sooner). [01:33] but apt installing something is easy [01:33] Goodbye, Webfaction [01:33] Onto rackspace [01:33] Good for me to have $700 out in the cloud [01:34] you are guaranteed to get $700 back, but it may not all be your $700. [01:34] Well, actually [01:34] You know what? [01:34] Fuck it [01:34] One more fucking day of fucking viruses on the fucking mediawiki. [01:35] Because fuck it, I don't need this shit. [01:35] I can handle the giggling bucktooth nerd feedback that we're "hacked" for 24 hours until I cool down [01:36] Rackspace doesn't look like a good fit. [01:38] SketchCow: sorry for the hassle. I seriously did not mean to provoke you into trying to fix this right away, I was just talking about it this afternoon because I was stymied that I seemed to be the only person afflicted. [01:39] what exactly happened [01:40] I run mediawiki too and I'm scared [01:41] do we just want a wiki? [01:41] if so, github has a great wiki service [01:42] powered by git, so we're not relying on them either [01:51] Yeah, no. [01:51] I'm speaking with the host for textfiles.com. He's a fan, he's good, he helps me. [01:51] We'll straighten this shit right out [01:52] What's up SketchCow [02:08] Three sniper bullets short of a high school shooting [02:08] hang in there, brother [02:08] if there's anything I can help with, let me know [02:09] I'm just going to go with the guy, I'm sure. [02:09] I like The Guy [02:10] Also, I am poor these days. [02:10] $300 for 5 years makes sense. [02:10] $240 a year? Not so much. [02:24] Cheered up. [02:25] Going with the textfiles.com host, he's been there for me over the years. [02:25] Also, the Musee' Mechanique just gave me permission to film in there. [02:25] No lights, no wires on floor, one guy, one camera. [02:25] ah, fun place [02:25] Yes! [02:25] I knew they likely got a lot of shit-ass crews. [02:25] that should be fun [02:25] liked it better when it was on the cliffs [02:25] that should be really awesome [02:26] I went there once and it was an awesome place [02:26] > Please cancel my "archiveteam" account effective immediately. I would like to take advantage of your 60 day money back guarantee and request a full refund. [02:26] I confirm that your account has been closed and a full refund has been sent. [02:26] > Your site is for nerds telling geeks how to act like robotos. [02:26] Can you tell us what exactly did you find hard/difficult in our panel? [02:26] [02:26] Did you consult our docs at docs.webfaction.com or try to contact support about it? [02:30] Note to self: never recommend anything to SketchCow. Ever. [02:32] many have tried / few have not died [02:32] I have this hilarious exchange [02:33] Where Tim Berners Lee is talking to people about whether he has the old webserver code [02:33] And he says not, and mails us at the archive for a copy [02:35] and the archive says... [02:35] Also, it's my fault, I should have known that the guy who has sex with Heroku servers and made it his full-time job would probably look at a web host that lauds the experience of steering a hay truck with bash scripts and think "oh, he's gotta get in on that delicious buffet" [02:36] Like, if you're not into spicy food and you ask the guy whose cube smells like you got punched in the face by Vindaloo where's a good place to get grub [02:36] The archive is at a loss, but thanks to the links, we'll be in better shape [02:40] * SketchCow kennethre NO GETTING BITTER [02:41] My quest in life is now to start a user-content company that gets massively popular and suddenly kill it without warning. [02:41] No, no, it doesn't taste right if you plan ahead [02:42] It's gotta be a surprise [02:42] haha [02:42] You have to think it's all for the good of humanity and speak at SXSW and in interviews [02:42] Then when you sit at your desk, head down, while they haul out the beanbags, you look up with seething eyes [02:42] changing the world [02:42] Soulant: cloud spirituality [02:42] I'm laughing so hard right now [02:43] I will crush your dreams as mine have been crushed [02:43] also sex with Heroku servers could be awesome [02:43] like [02:43] need more? [02:43] HEROKU SCALE [02:43] yeah [02:43] yipdw: it is indeed [02:43] They surround you with their little mouths [02:43] All want some shared you [02:44] It goes to 11. [02:44] that said, I deployed a CouchDB-based app to heroku + cloudant today [02:44] and after trying to hook up the app to the Cloudant DB I saw the "You're Doing It Wrong" talk that specifically mentions why couchrest_model is bad [02:44] guess what I'm using [02:44] so uh I don't know what the point of that was, except thanks Heroku, I guess [02:45] <3 [02:51] need more? [02:51] HEROKU SCALE [02:51] hahaha [02:51] if you're a pimp, you can even use heroku ps to keep track of all the bitches you've choked [02:52] Here, cheering up emo kennethre [02:52] http://www.archive.org/~jake/free-couch/ [02:52] Free Couches on Craigslist. [02:52] haha awesome [02:52] :D [02:52] wtf? [02:52] yes, but when can you deliver the couch? I live on the 4th floor. [02:53] not bad: http://www.archive.org/~jake/free-couch/5Id5L75W43K23F53J4c1ob627dceda7c51c7d.jpg [02:53] crazybeans. [02:58] http://i.qkme.me/36ckpl.jpg [03:05] OK, I'm going to call it a day [03:05] I got some stuff done, but I'm behind [03:05] Tomorrow, tomorrow things get done. [03:05] I'm moving URLTE.AM to my textfiles.com machine, it's a fucking static site and people have been very patient with me about it. [03:09] OK. [03:09] Over the next day or two, urlte.am should go to textfiles.com. When it's looking right, I'll plunk our replacement back. [03:11] DNS servers: [03:11] ns1.easydns.com [03:11] ns2.easydns.com [03:11] remote1.easydns.com [03:11] remote2.easydns.com [03:11] ns3.easydns.org [03:11] ns6.easydns.net [03:11] Well, that was fast. [03:13] OK, it'll be good shortly. [03:13] Good, something off the fucking plate [03:13] Now, to go see SF [03:13] If kennethre only LIVED here, we'd be having DINNER [03:13] enjoy your evening, brother [03:13] on a fucking plate [03:13] SketchCow: :( [03:14] SketchCow: I'll be in san jose next thursday [03:14] er, santa clara [03:19] some of you may have seen this already, but it's a really awesome archive thing, so I'd like to share it again: http://infovore.org/archives/2012/02/26/a-year-of-links/ [03:20] that is a good one [04:06] "Do you want to type them in by hand?" [04:07] shit... i wish i still had that "The Internet Yellow Pages" book [04:08] perhaps I still do have it in a box somewhere [04:09] metadata: apparently always my enemy [04:09] even such simple things as "what books are in this box?" [04:11] oh god, the Internet Yellow Pages [04:11] Michael Wolff, right? [04:14] i don't remember [04:17] looks like there were several books by that name [04:18] iirc, the cover looked like this: http://ecx.images-amazon.com/images/I/414R8NBC1XL._SL500_AA300_.gif [04:24] SketchCow: If you're still accepting images for your next talk, Mine is on the wiki: http://archiveteam.org/index.php?title=User:Scumola [04:27] Also, tweets mentioning 'oscar' during last night's broadcast (I've got to do something with all of that data): https://skitch.com/scumola/8fmex/steves-twitter-trend-grapher [04:28] "oscar 17 crystal 1"? [04:28] assuming billy crystal [04:29] but what is the number? [04:29] That just happens to be the values on the graph where my mouse was. Ignore that. :) [04:31] Also, I've been unwinding all urls from url shorteners that have appeared on the twitter spritzer feed since Oct, 2011. I've got about 23M unwound urls now and counting. [04:33] fuck yeah [04:33] urlteam++ [04:34] Yea, I haven't forgotten about urlteam. :) [04:37] I've got the whole twitter thing down to a science now. I have nagios watching the stream, so I get paged when the stream goes down or I've lost more than a couple minutes of data. I archive (bzip2) the data in minute-buckets and then go through each minute bucket and unwind all of the urls recursively and insert into a DB. I keep one week's worth of the full tweets in the DB also for [04:37] twitter-based searches if I need it. [04:37] I've got a small sqlite database of goo.gl short urls that showed up in a torrent listing site rss feed [04:38] how large do the minute buckets generally get? [04:38] Coderjoe: That's cool. I was initially scraping the url-shorteners too, but I figured that I wanted the ones that people were actually using. [04:39] well, mine is because I wanted to use the feed, but some people submit the torrentinfo page link on another site, while pretty much everything else is a link to download the .torrent file [04:39] each minute is only like 5MB or so uncompressed. Tweets are only about 2k each and the spritzer feed is like 2% of the full firehose. [04:40] but it's enough to get a good sample size and most urls are re-tweeted anyway, so I am probably capturing most of the 'important' urls. [04:40] each hour is about 350MB of data. [04:41] and some of them are behind goo.gl urls. so I wrote a php script that pulls the feed, unshortens goo.gl urls (and saves the results in a database for later lookup, so I don't hammer google each time), and then convert the torrentinfo links to download links [04:41] tweets are 2k? [04:41] a day is about 6.9GB uncompressed, 700MB compressed [04:41] Yea, thereabouts. It varies, but 2k is a good ballpark figure. [04:41] for a 140-char message? [04:42] that's a lot of overhead [04:42] yea, there's a bunch of metadata in there. Lat/Lon, user profile stuff, ... [04:42] 17:54:58 +0000 2011","profile_sidebar_fill_color":"ffffff","id":306331380,"utc_offset":-21600,"favourites_count":7,"url":null},"in_reply_to_screen_name":null,"truncated":false,"id":174353887023730688,"created_at":"Tue Feb 28 04:43:00 +0000 2012"} [04:42] i0.twimg.com\/profile_background_images\/414657846\/tumblr_lyqhl7NzZL1r972d6o1_500.png","profile_background_color":"EBEBEB","protected":false,"id_str":"306331380","profile_background_tile":false,"profile_image_url_https":"https:\/\/si0.twimg.com\/profile_images\/1847372264\/sel_normal.png","name":"AA\u2665","default_profile_image":false,"follow_request_sent":null,"created_at":"Fri May 27 [04:42] ription":"","is_translator":false,"show_all_inline_media":true,"following":null,"profile_background_image_url":"http:\/\/a2.twimg.com\/profile_background_images\/414657846\/tumblr_lyqhl7NzZL1r972d6o1_500.png","default_profile":false,"profile_link_color":"b6ffb6","time_zone":"Central Time (US & Canada)","verified":false,"geo_enabled":false,"profile_background_image_url_https":"https:\/\/s [04:42] ser":{"notifications":null,"friends_count":235,"profile_sidebar_border_color":"DFDFDF","screen_name":"andy_arceo","contributors_enabled":false,"lang":"es","statuses_count":28438,"profile_use_background_image":true,"location":"","listed_count":0,"profile_text_color":"000000","followers_count":203,"profile_image_url":"http:\/\/a0.twimg.com\/profile_images\/1847372264\/sel_normal.png","desc [04:42] {"contributors":null,"text":"y pulido me aplico la flechita OTRA VEZ","entities":{"urls":[],"user_mentions":[],"hashtags":[]},"in_reply_to_user_id":null,"place":null,"retweeted":false,"coordinates":null,"retweet_count":0,"source":"web","in_reply_to_status_id_str":null,"geo":null,"in_reply_to_status_id":null,"favorited":false,"in_reply_to_user_id_str":null,"id_str":"174353887023730688","u [04:43] That's one tweet. [04:43] yum [04:43] so the full feed is how much? [04:43] the full feed costs money, but is about 70Mbps uncompressed. [04:43] whee [04:43] the spritzer feed (the one that I get) is free and it's only 1.5Mbps uncompressed. [04:44] I'd love to get my hands on the other social media feeds like foursquare and others. [04:44] i bet we could get twitter to send us dumps [04:45] i know some people in ops over there [04:45] they charge money for the data - big-time. [04:45] only the sprtizer feed is free. [04:45] yeah but we're archivists, it's different [04:45] plus, the full twitter feed of everything would be huge, I'm guessing. [04:45] compressed, it shouldn't be SO bad [04:46] statuses/firehose [04:46] Returns all public statuses. The Firehose is not a generally available resource. Few applications require this level of access. Creative use of a combination of other resources and various access levels can satisfy nearly every application use case. [04:46] heavily compressed [04:46] oh nice [04:47] Returns a random sample of all public statuses. The default access level, ‘Spritzer’ provides a small proportion of the Firehose, very roughly, 1% of all public statuses. The “Gardenhose” access level provides a proportion more suitable for data mining and research applications that desire a larger proportion to be statistically significant sample. Currently Gardenhose returns, very roughly, 10% of all public st [04:47] statuses/sample [04:47] atuses. Note that these proportions are subject to unannounced adjustment as traffic volume varies. [04:48] Just the spritzer feed (2% of the full firehose) is (on my system) ~90GB compressed for everything since Oct 2011. [04:49] i wonder what level the randomization happens [04:50] like, would different connections get different results? [04:51] I think that each "hose" is just a modulus of the tweet id. So, the spritzer is $id mod 100 = 1 or something like that. [04:51] so everyone getting the spritzer feed would get the same tweets. [05:19] damn, 90GB of tweets [05:20] I hope someone has wordclouded that shit [05:20] yipdw, from who? [05:21] Zuu-: swebb2's copy of the spritzer feed [05:21] ah [05:46] hai, is there a channel for mobileme? [05:46] #memac [07:40] Pardon my ignorance, but is archiving Twitter necessary, given the LoC archive? [07:41] swebb2: loc? [07:41] oh nvm duh [07:41] Library of Congress [07:41] yeah [07:44] Is it to provide a public dump vs. request only? [08:21] * joepie91 is running a few boxes on the mobileme thing [08:21] er [08:21] * joepie91 is running the mobileme thing on a few boxes * [08:22] Jump over to #memac, that's where the mobileme party is going [08:22] o [08:23] ^_^ [08:40] Hi, [08:40] Psst. [08:40] I've talked to people. [08:40] Inside people. [08:41] The LoC tweet archive? [08:41] I wouldn't buy a lot of stock in it. [08:41] Wouldn't bet the farm, you know. [08:43] Wouldn't bet the farm, you know.Juuuust sayin' [08:44] figures [08:47] Yeah. [08:47] So do whatever we can [08:47] Oh, that new wikimedia project. Daddy likes. [09:02] FUCKING FINALLY. [09:02] URLTE.AM IS BACK [09:02] SketchCow: thank you! [09:02] oh, not me. [09:02] Now, let's not have THAT happen again. [09:02] What are you up to, radii? [09:03] I've been saying "Free the edit stream!" to every wikipedian I come across for years, so when I saw the WM Commons project I figured I should get online [09:03] Ah, OK. [09:03] Welcome aboard [09:03] URLTE.AM finally back. [09:04] but alas I have a timezone and dayjob challenge, so I'm just going to idle here for the next long while. [09:04] We're exactly like Wikipedia is in how we run things [09:04] what, no, deletionists are not welcome here. [09:05] but with less drama, less deletes and a lot more rape [09:05] no deletes. [09:05] no is less! [09:05] * chronomex shakes head [09:08] * radii deletes my awakeness [09:08] * ersi injects a coffee machine into radii [09:08] ersi: in 8 hours I will agree with that. [09:11] http://web.me.com/stressakustik [09:11] awesome page [09:20] Tomorrow, we see about the wiki. [09:21] Oh, I will love THAT overlap. [09:38] damn.. while working 12 hour days kicks ass on payday, the rest of the time it sucks [09:48] dcmorton: I was just reading a blog post about that: http://thecodist.com/article/why_i_don_39_t_do_unpaid_overtime_and_neither_should_you [10:00] worth a read: http://www.stevepavlina.com/blog/2006/07/10-reasons-you-should-never-get-a-job/ [10:08] http://qaa.ath.cx/LoseThos.html [10:09] ^ i thought the post was jumping the shark, but if you take the time to watch the videos and look at his twitter stream, I think the other's quite correct [10:14] joepie91: that post is a bit rediculous in parts [10:14] kennethre: it's hyperbole, obviously :P [10:15] joepie91: this guy must have had a really soul-crushing gig :) [10:15] nah [10:15] imo hyperbole is necessary there [10:15] for the masses I'd agree [10:15] it throws people out of their normal line of thinking and makes them reconsider :) [10:16] yes, exactly [10:16] those that don't need it will recognize it as being hyperbole, usually :) [10:16] I used to do hyperbole posts [10:16] I feel like I grew out of it though [10:16] also, lookie at http://www.ramhost.us/?page=status-core [10:16] if you didn't read #memac yet [10:16] :P [10:16] the bottom two graphs [10:16] geeze [10:16] haha [10:16] * joepie91 enjoys seeing those graphs pealk [10:16] peak* [10:16] I just briefly hit 156mbit total [10:17] before that I was on 130mbit total for a while [10:17] 60mbit down and 70mbit up or so [10:17] I do have the idea I'm going to be running out of bandwidth very soon though.. ._. [10:17] or traffic*, rather [10:17] maybe they only limit incoming traffic? [10:18] 88.211 GB of 600 GB (14.7%) [10:18] oops [10:18] :x [10:18] that was on 3GB a few hours ago [10:18] per box or per account? [10:18] that's that one server [10:19] I have 5 other servers, but I can afford for this one to be cut off due to bandwidth exceeding [10:19] kill it and start a new one when you're doing :) [10:19] *you're done [10:19] :P [10:19] it's just a storage server [10:19] oh man [10:19] doesn't really do much [10:19] i think i used to use these guys [10:19] I did! [10:19] ramhost? [10:19] they're <3 [10:19] no they were horrible [10:19] worst support ever [10:19] you sure you're talking about ramhost? [10:19] pretty sure [10:20] i don't know of any others that offer both openvz and kvm [10:20] buyvm does [10:20] though it was openvz and xen at the time [10:20] and so does some .nz provider [10:20] mabye it was someone else [10:20] nah [10:20] ramhost never did xen [10:20] hmm [10:20] * kennethre searches the logs [10:20] ah [10:20] vpslink.com [10:20] so, so terrible [10:21] ramhost always scores really high on support, the one issue some people have with them is that they're rather straightforward and to the point [10:21] :P [10:21] and some people experience that as 'unfriendly' [10:21] mm [10:21] never heard of them [10:21] OH [10:21] this reminds me [10:21] I have an unmetered santrex box laying around [10:21] not as bad as liquidweb though [10:21] no one's as bad as them [10:21] might as well make it work on this as well [10:21] theoretically that box was terminated in december [10:21] >.> [10:22] according to their administration, that vps doesn't exist anymore [10:22] according to my ssh client, it does [10:23] and I've heard mixed reviews about liquidweb [10:23] and I'm pretty sure that santrex is worse than liquidweb :P [10:23] far worse [10:23] doubt it [10:23] haha [10:23] well [10:23] It's amazing how many interpret straight forwardness as unfriendly [10:23] let me make a list :P [10:23] Or being a douche [10:23] I ordered a santrex vps [10:23] I'm usually straight forward [10:23] after the billing failed and had to be done manually [10:23] it assigned the wrong amount of RAM [10:23] to my vps [10:23] classy [10:23] 24 hours after that it went offline for 2 hours [10:23] after which the RAM suddenly fixed itself [10:24] however, while I picked debian, it had installed centos [10:24] they also did not send me an email where to find my solusvm panel [10:24] so I had to google for the address to figure out how to reinstall my box [10:24] as the santrex panel itself did not allow for doing that and only showed basic stats [10:24] that was the first 24 hours after placing my order. [10:25] I had an uptime monitor checking my site [10:25] not a day has gone by where I did *not* get a downtime alert [10:25] hahaha [10:25] for months, every day it would go down multiple times [10:25] mine was worse [10:25] *not worse [10:25] for minutes to hours [10:25] now comes the best part [10:25] they told me they would 'migrate to new hardware and renumber' [10:25] "somewhere next week" [10:25] no date or time was given [10:25] and in the end, it was suddenly offline for two days somewhere at the end of that week [10:26] without an announcement beforehand with a more specific time [10:26] after two days it was back up with a new IP [10:26] what they did *not* tell me, was that, besides upgrading hardware and renumberin, they also moved my server to an *entirely different datacenter* [10:26] which messed up the redundancy of my storage grid to a degree [10:26] ah, hosting [10:26] they moved it from ecatel to leaseweb [10:27] ah that is kind of what happened to me [10:27] after that I had constant issues with downtimes [10:27] i wanted a refund for various reasons [10:27] a few times it went down for a few days [10:27] and they decided to give mea dedicated box instead from some other company [10:27] without any further announcement [10:27] that is run by the same people [10:27] joepie91: forgot to pay the electric bill [10:27] sending a support ticket would just return "look at our status page " ( which, I might add, had no reference to the outage) [10:27] joepie91: two more user payments and we can be back online! [10:27] at one point I suspect the server was seized, when it was mysteriously offline for 5 days [10:27] and lol [10:28] and then I missed an invoice, and my server was 'terminated' [10:28] and till the day of today it's still running [10:28] taken out back and shot? [10:28] 2 of my 3 IPs have been renumbered to another vps [10:28] reassigned* [10:28] the third still works [10:28] lol [10:28] so yeah [10:28] santrex. [10:28] don't touch it with a ten feet pole. [10:28] I'm pretty sure that was worse than liquidweb :P [10:28] i have a friend that worked there [10:28] and it's apparently run by like the greediest person in the world [10:29] yes, I know a friend of the owner and he said the exact same thing [10:29] or well, 'know' is a big word [10:29] but I occasionally talk to him [10:29] hahaha [10:29] what'd they say? [10:30] pretty much that it's someone that doesn't really give a shit about whether servers are running [10:30] as long as the invoices are paid [10:30] (by customers) [10:30] since I don't know the owner personally I obviously don't know how true that is [10:30] :P [10:30] but it would not surprise me at all [10:31] pretty much exactly what i heard too [10:31] some other fun ones: [10:31] w2servers and veeble [10:32] w2servers was one of lyron foster's projects [10:32] you might have heard of him [10:32] to get an idea: http://www.google.nl/webhp?sourceid=chrome-instant&ix=seb&ie=UTF-8&ion=1#hl=nl&safe=off&output=search&sclient=psy-ab&q=w2servers&pbx=1&oq=&aq=&aqi=&aql=&gs_sm=&gs_upl=&fp=d60c0cceab75dfa0&ix=seb&ion=1&bav=on.2,or.r_gc.r_pw.r_cp.r_qf.,cf.osb&biw=1366&bih=680 [10:32] oh fun [10:33] I had one of those $24/yr servers there as a throwaway server [10:34] man hosting is shady [10:34] i hate how that's always the case too [10:34] and then there's veeble, which is, without malicious intent, just ran by someone that doesn't have a clue [10:35] soooo [10:35] let's get that santrex box archiving [10:36] joepie91: ah he worked for hostgator [10:36] joepie91: that explains everything [10:36] hahaha [10:36] even hostgator isn't /that/ bad, though :P [10:36] shared hosting always sucks [10:37] hmm, I have to disagree [10:37] the only place i've found that to not be true is webfaction [10:37] i'm sure there are others [10:37] but they're very rare [10:37] I have shared hosting for my personal site at a dutch hosting company [10:37] webruimtehosting [10:37] sounds incredibly scammy, but they're very good [10:37] it's like 24 euro a year with free domain included, and their server is *fast* [10:38] and usually 6 out of 7 servers hit 100% uptime each year [10:38] the other is usually 99,99% or so [10:38] I've been rather impressed with them :P [10:38] you get what you pay for :) [10:38] heh, never in history has such a widespread saying been disproven so thoroughlu [10:38] thoroughly* [10:40] * joepie91 wonders how much the santrex box will push in terms of bandwidth [10:41] ooo, cinfu box is now hitting 80mbit :D [10:41] * joepie91 is enjoying looking at the terminal-based blinkenlights [10:47] ... oops, I just took down the wrong network interface on my santrex box ._. [10:47] that was stupid [10:48] that was stupid. [10:48] um, anyone has any idea how to fix this? :3 [10:54] ouch. when Pandora doesn't like your browser, they really lock you out [10:54] you can't even find out what Pandora _is_ [10:56] db48x: their loss? :P [10:58] yea, I've got other music I can listen to [10:58] for Free music: http://www.jamendo.com/ [10:58] for everything else: http://www.vk.com [10:58] indeed [10:59] preferably jamendo, ofc :) [10:59] I'm listening to http://www.youtube.com/watch?v=ME5urFBf0kk&feature=youtu.be&hd=1 at the moment [14:27] sketchcow [14:31] He's here and reading/responsing occationally [14:31] what's up? [14:32] i just came [14:32] to thank this ragtag team of yours [14:33] ah, hehe [14:34] on behalf of what ever little I've contributed I'll say you're very welcome [14:44] ersi [14:44] how do i send any sites i have done myself? [14:48] Good question, dunno if there's a good answer - besides possibly making a upload to the Internet Archive or making an archive(zip,rar,tar,etc) available somewhere (http/torrent etc) [14:48] where do the archiveteam keep their dumps?" [14:49] we're a ragtag team like you said [14:49] so no central storage [14:49] we share stuff between ourselfs or make them public in other ways [14:49] i guess that works, for now [14:50] ersi, try using osiris SPS [14:50] check it out now [14:50] in some cases, Internet Archive swallows whatever we produce after it has been made nice [14:51] FAMAS|2: also have a look at http://retroshare.sf.net/ [14:52] Or FreeNet, or tahoe-lafs.. or.. or.. [14:52] i got that thing yesterday from anonops [14:52] :P [14:52] actually tahoe-lafs is something entirely different [14:52] ersi, check osiris SPS [14:52] it may blow your head off [14:52] tahoe-lafs is storage, not sharing/collaboration/communication [14:52] git-annex [14:53] I don't get turned on by that, so I doubt it FAMAS|2 [14:53] joepie91, why not use the existing irc network to share? [14:53] FAMAS|2: what do you mean? [14:53] "sharing/collaboration/communication" [14:53] you can send and recieve file via irc [14:53] yes... [14:53] which is centralized... [14:53] ersi, let me get a tutorial for osiris [14:53] whatever we are doing now works pretty well [14:53] FAMAS|2: stop, I don't care [14:54] the whole point of retroshare, osiris, etc, is that it's decentralized [14:54] if you want to do something with any data someone here produces, do it [14:54] if you feel like doing something, do it [14:55] http://www.metacafe.com/watch/yt-0eNP3RT7rmY/osiris_sps_tutorial_create_a_portal/ [15:18] FAMAS|2: http://www.archive.org/details/archiveteam [15:21] ugh, that smf forum mirror downloaded 8GB of which only 2.4G remained because of the not-saving the *msg* clutter [15:36] ndcu ROCKS [15:40] are there file extension larger than 4 chars? [15:41] .torrent [15:41] .vbproj [15:41] lol [15:42] .aria2c [15:42] Plenty [15:43] But many people are still afraid of long file extensions [15:43] and extensions with more than 1 dot? [15:43] Technically no, but there's .tar.gz [15:43] my.random.text.files.txt [15:43] Which is really a tape archive which is then compressed with gzip [15:44] It's modular. You can compress anything with gzip like that [15:44] .sql.gz [15:44] Whatever [15:44] So, it's really two independent file extensions, is what I'm saying [15:45] k tahnks [15:46] klol np [15:49] wow, smf is retarted [15:49] #klol is awaiting for some care [15:49] eg on &topic=6461.0 there are links to the prev and next thread [15:50] their targets are &topic=6461.0;prev_next=prev and &topic=6461.0;prev_next=next [15:50] and yes, you end up at those URLs for topics that actually have proper unique different IDs, eg 6460 and 6462 (examples, not real) [15:50] idiotic [15:50] what defines ordering? most forums I've used reorder threads based on activity [15:51] Schbirid: What if thread 6460 gets deleted between the time the user opens the page and clicks the link... [15:51] Coderjoe: ah, that might be why i have much different IDs as prev and next (eg 2523) [15:52] wow, phpbb2 does it too [15:53] nitro2k01: bad luck? what is the difference? [15:53] i get the impression that doing SMF with wget --mirror will be a pain. perhaps a special toolset? [15:53] i am writing my experiences [15:54] it does? I could have sworn that phpbb actually used params like &start=30 for page three, &start=45 for page four, etc (given a posts-per-page of 15, at least) [15:55] It may be annoying for you, but it makes sense to have those links [15:55] (and where I mean phpbb2 when I say phpbb) [15:56] nitro2k01: how? [15:56] example here http://rome.ro/smf/index.php?topic=6204.0 [15:57] not pagination of a single topic, i mean the previous/next topic [15:57] Oh wow [15:57] I thought it would redirect [15:58] THIS doesn't make sense [15:58] oh. I generally never use those [15:58] when I do, it is accidental and I meant prev/next page of the current thread [15:59] "I don't always use prev/next, but when I do, it's accidental" [15:59] stay frosty, my friends [16:02] step 1: 7zipping all i got before i start decluttering [16:06] uhoh. I fear the size of this mobileme profile... username is "emusicbox" [16:07] heh [16:07] feed.xml is taking awhile [16:09] oh great smf/index.php/topic,1998.0/images/smiles/icon_smile.gif.html [16:09] yay. leaving the default text/html content-type header [16:10] I wish browsers were a lot more strict on the handling of content-type [16:10] forcing website code monkeys to return the correct type [16:11] - Running wget --mirror (at least 18934 files)... [16:15] Why would you want to go to the next thread, that's tarded.. what is the index/subforum index for [16:15] oh yes, that idiocy was implied :D [16:15] actually it might be useful if that function only shows unread topics to registered users [16:16] imo it's retarded [16:16] larger than a small group of people on a forum makes threads "too diverse" to serve everyones interests [16:16] it also serves the forum index for 404s [16:16] lol [16:16] what can I say.. forum software.. [16:17] to be fair, this is 4 years old [16:18] and the web is 20 years old. how to handle 404s should be well known by now [16:22] wait, that was incorrect [16:23] it does serve 404 for nonexisting files [16:23] but not for eg "this thread does not exist" [16:42] i had to write a workaround for 4chan for awhile, as some servers were using versions of the board code that didn't set the 404 status code [17:25] SketchCow: Ah, yikes; I knew LoC was behind the curve, but I figured Twitter was in their ability to save [17:26] (And how often is metadata@textfiles.com checked?) [18:22] to be archived? http://www.openimages.eu/media [19:17] I am metadata@textfiles.com. [19:17] I have a tad of a backfill. [19:18] Are youu Alex? [19:19] Also, everyone will be delighted to know the leak of the JSTOR liberator caused JSTOR to shit and get off the pot. [19:19] They're now arranging to send all their stuff to archive.org (the early stuff) [19:20] so who leaked it? [19:20] I don't know, honestly. [19:20] But the fact that it went to the circlejerk hostname does not make me think it's anyone new [19:52] our logs are public, yes? [19:52] y [19:52] CANDY VAN [19:52] mmmkandi [19:52] http://badcheese.com/~steve/atlogs/?chan=archiveteam [19:53] I always forget it [19:56] #archiveteam is forever [19:58] Where is all the data stored? ;p [20:00] iiiinternetttt [20:10] INTERNETNENTTTT [20:10] I just did an interview with NZ tech podcast. [20:10] PROBABLY was a little radical sounding [20:12] Did you bring it up to 11? [20:14] I said we built the internet like a car with no brakes and now we're retrofitting brakes as it careens down the hill [20:15] it's agile development [20:15] nobody had a use case for brakes yet. [20:16] YOU WOULDN'T DOWNLOAD A CAR [20:16] I would download so many cars [20:16] hehe [20:17] I'M DOWNLOADING A CAR NOW [20:17] http://bitcoinmedia.com/uploads/2012/01/You-Wouldnt-Download-A-Car.jpg [20:17] when mobile broadband takes off in cars, I can download a car from my car [20:17] I used a variation of that image for a talk from two years ago [20:19] YOU WOULDN'T STEAL A CAR [20:20] YOU WOULDN'T SHOOT A COP [20:20] AND STEAL HIS HELMET [20:20] * Schbirid poops in it [20:20] TWICE [20:20] YOU ONLY HAVE ONE HEAD AFTERALL [20:20] DON'T OVERSHOOT COPS [20:20] otherwise you won't get the pacifist run achievement [20:20] http://blogs.loc.gov/digitalpreservation/2012/02/jason-scott-rogue-archivist/ [20:21] I thought you described yourself as a free-range archivist [20:21] are there more photos of your LOC talk around? [20:22] I didn't talk at LOC yet [20:22] This is an interview [20:23] I did a PDA talk [20:23] This is just e-mail [20:23] The photo is of me at ROFLsummit [20:23] whoa, totally mixed up your pda talk with being the loc [20:32] SketchCow: Yes, I am, and don't sweat it; you seem insanely busy [20:33] I was just curious if you still checked it at all [20:37] I will be assigning you stuff. [20:37] Shortly [20:37] Thanks [20:37] Today is catchup day [20:37] Like I said, put out your fires first [20:37] A day we call HOW THE FUCK DOES JASON GET SO MUCH DONE [20:38] before heading in to work today, I began the process of bringing together all my stage6 videos. What should I do about the metadata, which is currently in mysql? just a mysql dump, or should I parse it and output xml or something? [20:39] and how should I package it for IA? [20:39] stats (including some about file sizes) at http://wegetsignal.org/stage6.php [20:46] hmm [20:46] what was that github grabber? [20:47] balrog: https://github.com/joeyh/github-backup [20:47] Schbirid: that pulls all branches, right? [20:47] no clue [20:47] hrm, ok [20:48] Schbirid: yes [20:48] also did he fix that 30-of-each-thing bug? [22:02] I am up for a .txt of the metadata that I can convert [22:28] Who wants an easy scrape project? [22:28] Might be less than 15 minutes of work. [22:28] http://www.1up.com/do/minisite?cId=3148397 [22:28] I want a textfile with the date, and then the description, in a textfile. [22:31] One textfile per, or one with all of them? [22:35] Let's go for a textfile with the date in the form of MM-DD-YYYY followed by a one line description ripped from those pages. For the GFW entries only. [22:35] I show it'll be about 130-140 lines. [22:35] I can then add the full podcast to archive.org. [22:35] Gotcha [22:44] Why that date format? [22:45] Let me see some ISO8601 up in this bitch [22:46] MM-DD-YYYY is still ambiguous as far as MM and DD go [22:46] Still better than having everything as two digits each, but hey... [22:50] YYYY-MM-DD > MM-DD-YYYY [22:52] and having dropped my pearl of wisdom for the day I'm off for home. see y'all later. [22:55] a .txt of the metadata for each video? alongside the video? [22:55] That date format is ebcause that's what the filenames are, bitch [22:55] It'll go to the other format for the date [22:55] I'll make it all good [22:57] (my question is about my stage6 data) [22:58] the metadata I have is uploader, tags, stage6 video ID, original url, title, description, filesize, and possibly date of upload. (not sure if this field was upload date or date I added it to the queue, offhand) [23:00] I also have more metadata than videos [23:01] That's excellent, actually. [23:01] Yes. [23:02] Want me to give you a place to upload it or do you want to do it. [23:03] I can do it, I think with some guidance on things like number of videos per item and preferred metadata info format [23:03] or I can just organize it a bit and upload it for you to ram into IA [23:05] yeah. looks like the v_added field is the date that stage6 said it was added to the site [23:05] but I only have that for videos I managed to scrape, I think [23:06] which appears to be all the videos I indexed (which was sadly not every video on the site) [23:06] What kind of videos are these? [23:07] technically, AVI files (they're divx's .divx by filename) [23:07] divx networks ran a user video hosting site for awhile, named stage6 [23:07] Yes [23:07] Just tracked. [23:07] How many videos. How big. [23:08] This is 50% curious, 50% useful for advising you. [23:08] though some of these are probably tv shows [23:08] easy stats are here: http://wegetsignal.org/stage6.php [23:09] I have 4989 actual downloaded videos, ranging from 556K to 1.9G [23:09] I see. [23:09] Before it died. [23:09] So 303gb [23:09] yeah [23:11] Just talked here. [23:11] One item per video [23:11] SketchCow: Done [23:11] alright. current file names are like 1.divx for video id 1. should I rename them at all? [23:11] Let's do one. [23:11] Choose something fucking hilarious. [23:12] Upload it, plug in the metadata. [23:12] Give me the item, I'll make an archiveteam-stage6 collection with it [23:12] And we'll go from there, OK? [23:13] shaqfu: mail that in. jason@textfiles.com [23:14] alright [23:16] Sent [23:29] Hmm, not here yet. [23:29] Odd [23:29] So, shaqfu: What interests you: general computer history items, or arcade manuals? [23:29] SketchCow: What's there more a pressing need for? [23:29] And it's probably my old school's email server being awful again [23:30] PERSONALLY, I think the general computer history items are more useful. The arcade ones are more "fun". [23:30] I'll tackle the history items, then [23:30] In case it interests anyone: http://www.flickr.com/photos/gameboygenius/sets/72157629470674857/ [23:30] Will do a writeup on it later [23:30] If push comes to shove, you could probably open the arcade manuals to crowdsourcing or w/e, since it's higher interest [23:30] nitro2k01: O_O that is amazing [23:31] I've already opened it generally [23:31] So yeah, bitsavers is the way. [23:31] Any collections you have in mind? [23:32] Oh, let me help. [23:32] The collection I'm putting up: [23:32] http://www.archive.org/details/bitsavers [23:32] As you see, 44 items. [23:32] Gotcha [23:33] http://www.textfiles.com/bitsavers/pdf/ [23:33] Now, that's the source area. [23:33] If you choose anything, I prefer you "close out" a directory. [23:33] I want the URL, and the date (as much as you can guess), and a desc. [23:34] Do as many as you can stand. [23:34] Sounds good [23:34] I'll be able to add them. [23:34] Do you want subject keywords also? [23:34] With only 44 of the thousands up, it's lots of space [23:34] I would like that but I don't want you murdered [23:35] SketchCow: I developed a pain tolerance to this sort of thing in grad school [23:35] I'll start tonight [23:35] OK, great. [23:35] I'll ensure the front page has credit. [23:35] One file per dir, or per item? [23:35] One file per item [23:35] Got it [23:35] I have a script that does the uploading, and lets me plug in your textfile. [23:36] Is it picky about formatting date/desc/keywords? [23:36] Not really. [23:36] Awesome [23:36] What you do saves me the hardest part. [23:36] I do the rest. [23:36] got it [23:37] I gotta run now; I'll start on this in a bit [23:37] Thanks for the work [23:38] Great wrk, nitro2k01 [23:44] SketchCow: what i meant about the wiki link for the networking film, was that there are still some perple who have yet to be ID'd, is all. [23:44] Ohhh [23:50] Archive Team wki is only about 600mb. [23:50] I'm getting our new host setup for it. [23:50] Then we'll move there.