[00:39] slow mobile me is still slow [01:54] I can.t put my finger on the precise Reasons. [01:54] I think the Audio-Quality needs work. [01:54] I hear a bit of echo near the start. [01:54] The Face (especially the eyes) is in some shots slightly out of Focus. [01:54] The Moving shot is a bit of a risk with fixed lenses. [01:54] for mug shots [01:54] The Focussing Bit near the End distracts a bit (I think Camera Lenses overshoot to fast when adjusting). [01:54] Barely literate critics, have to love them. [02:10] hahaha [02:19] Tell him that certain types of eyes simply absorb vast amounts of light into their cones, throwing the shot slightly out of focus.. it's rare, it's unavoidable, shit happens. [02:19] lol [02:20] "camera lenses overshoot to fast when adjusting" [02:20] lenses [02:20] what [02:22] eesh, yikes [02:22] [ec2-user@ip-10-243-119-16 files.splinder.com]$ pwd; ls -1 | wc -l [02:22] 22046 [02:34] * SketchCow boots zetathust, and does this: http://www.youtube.com/watch?v=Mu71EAdnjQ0 [02:43] if someone's got a better place for me to upload the 7z tell me, otherwise here's a link to it on mediafire: http://www.mediafire.com/?49kgs4umrb79a34 [02:44] If you get a dropbox you can upload it there and copy a public url [02:44] I don't actually [02:45] I don't really have anywhere else to put it online that I can share from, so my apologies there- but it is a pretty small download [02:58] Why not just throw on batcave? [02:58] I can make it browsable. [03:02] I don't have any logins or access- if you want to PM me something, I can throw it up there right away [03:03] you don't have rsync? [03:05] I do have rsync [03:10] so how would I go out using rsync to get the folder onto batcave? [03:15] okay- it's uploading to batcave [03:19] okay- it's up there [03:23] just as a note- there are some gaps in the archive, because some pages the site points to simply aren't there anymore [03:33] anything for the ffnet scrape [03:36] is it possible to upload directly into ia, using ftp [03:47] no, but you can use http [03:47] http://www.archive.org/help/abouts3.txt [03:47] oh, that's awesome [03:47] I didn't know IA had an S3 interface [03:48] that means I can reuse AWS::S3 and all the fun related bits [03:48] yeah, it's super rad. [03:49] that, and being able to specify metadata with http headers, means you can drop items into archive.org from shellscripts with 0 hassle [03:49] I quite like that [03:51] http://www.archive.org/create.php?ftp=1 [03:53] oh, yeah, you can do that but it's kind of lousy. [03:58] yes but its easier to do that , than to use ftp to login firsat and create the mxml by hand [03:58] btw, as a library, IA kicks LoC in the nutsack [03:59] loc's catalog is really nice [04:01] mostly because, when I search for anything in the Archvie, i can take the url of the *search* and dump it into jdownloader, which will then proceed to load and look for links, and find *every single result on the page*,and give me human readable links for them so i can poick and choose, without even having o click on each individual result [04:02] seriously thought, LoC website search is worse than useless, because it makes me give up, rather than keep looking, its just that bad [04:03] yeah the interface sucks [04:03] but at least the metadata is correct and somewhat consistent [04:03] yipdw: Unless you need to create items with directories [04:03] Then it sucks [04:03] Although, it's a lot easier now that I have internal access [04:04] oh, I was thinking of using it to shove WARCs at the IA [04:05] Then it's probably perfect [04:07] TECHNICALLY it's not S3 [04:07] It's S3 like. [04:08] Until this calms down: http://www.archive.org/~tracey/mrtg/derivesg.html [04:08] I'll be focusing on other things. [04:10] Oh wow [04:11] It's mostly ximm with all his forever-running heritrix crawls [04:14] yipdw: but wouldn't you need to hand vreate an xml for each warc file? [04:15] bsmith093: why would I need to hand-create it [04:16] S3 automatically creates the necessary XML based off of the headers you pass in [05:13] http://www.poe-news.com/forums/sp.php?pi=1002546492 [05:13] poe-news.com has announced they're shutting down. [05:14] start the warc [05:19] this good? wget-warc -mpke robots=off -U "Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)" --warc-cdx --warc-file=poe-news.com_12022011 www.poe-news.com [05:20] bsmith093: can you give me a very succinct idea of the current state of ffnet project? [05:20] or a very meandering, sloppy narrative [05:20] that'll work too [05:20] have ideas, cant code, got someihing half baked and don ish [05:21] underscor's working on a script to grab reviews and stories with storyinator [05:21] im just iterating thriugh every possible ffnet id, and culling the bad ones to make a linklist [05:28] underscor's way is almost certianly faster [05:32] spidering the site like yipdw suggested might be the fastest [05:33] arrith: can you explain #2 in the "extra credit"? [05:33] http://learnpythonthehardway.org/book/ex8.html [05:35] dnova: notice where double quotes get used versus where single quotes get used [05:35] there's something unique about he double quoted sentence [05:35] OH. [05:35] the single quote wasn't escaped [05:36] kinda [05:36] just that there is a single quote [05:36] usually when there's a single quote people use doubles [05:36] hmph. well ok. thanks :) [05:36] but yeah, you can escape it [05:37] i dunno actually if people usually escape or not [05:37] i've only seen doubles used then but i've only seen tutorialish code [05:56] spidering, i dont know how to tell wget to spider and save a linklist to then go back to [05:56] not spider with wget [05:56] spider with a ruby script that goes through the categories [05:56] also, on IA is it possible to edit an existing item? [05:56] or python script [05:59] wheres the script, and how do i runit? [06:00] ive got something by underscor from a repo, that looks like ruby [06:01] there isn't one [06:01] you gotta make it [06:01] ugh [06:02] pardon me by yipdw git://gist.github.com/1432483.git [06:10] eh? [06:10] oh [06:10] yeah hiws that going, any updates [06:10] yeah, I maintain that only hitting what you need to hit is the fastest way to do it [06:10] I haven't touched it since then [06:10] other work, etc. [06:11] I think arrith wanted to port it to Python [06:11] you can run it right now, if you have a Ruby 1.9 environment with the connection_pool, girl_friday and mechanize gems installed [06:12] ok , wonderful, now, how do i get those modules installed? [06:14] rubygems1.9.1 or 1.9 [06:19] haha [06:20] yipdw: yeah i was basically waiting to see what underscor ends up with and go from there [06:20] seriously how do i get those ruby modules installed? [06:20] possibly switching to a spidering method to get updates [06:20] bsmith093: http://www.google.com/search?q=rubygems+ubuntu [06:31] good god [06:31] bsmith: how many stories are on ffnet? [06:31] do we know? [06:33] less than or equal to 10,000,000 it looks like? [06:35] or: what is the highest valid ID you've found? [06:36] ~7million [06:36] can some kind [erson walk me through how to insatll girl_friday gem, ive found the darn thing but it wont install with gem install [06:37] https://github.com/mperham/girl_friday.git [06:41] anyone? [06:43] I have no ruby experience, sorry. [06:44] arrith [06:46] bsmith, [06:46] I think you need to relax just a little bit [06:48] I added the project to the wiki frontpage [06:49] yeah , i know, im overtired and really need to sleep [06:49] dnova: spot on. [06:50] ooh, thanks, chronomex [06:50] any ideas/critiques are welcome [06:51] I meant with respect to relaxing, but the link looks good :) [06:51] oh, lol [06:52] dang, it's been two months since I've uploaded anything [06:52] get busy time [06:53] are you running the fix-dld script or what [06:53] where are you getting all those splinder profiles!! [06:53] me? [06:53] I'm fix-dld [06:53] ahh figured :D [06:53] was offline for a while. [06:53] I'm downloading 2 users. have been for like 4 days [06:53] one is over 12gb [06:53] one is over 3 [06:54] I lost one that was over 10gb because I ran out of ram+swap :( [06:54] using tmpfs? [06:54] tmpfs is only a good idea for when you're doing a bunch of threads simultaneously [06:54] not the way its supposed to be (i.e., not a ramdisk) [06:55] ? [06:55] no, the upload I'm doing now is to archive.org and not an archiveteam thing. [06:56] http://www.archive.org/details/bellsystem_PK-1C901-01 [06:56] well, gnight/gmorning ,all, im gonna go sleep like i should have done 2hrs ago bye [06:56] bsmith093: sleep well. [06:56] sleep well! [06:56] arrrgh [06:56] :D [06:57] chronomex: ook now what? [06:57] bsmith093: ? [06:57] you said aargh [06:57] nvm [06:58] k night bye [06:59] heh. [07:07] bsmith093: easiest way to install it is to get a Ruby environment, get Bundler (gem install bundler), and then install all the gems in the bundle (bundle install) [07:31] Ops, please [16:42] http://rbelmont.mameworld.info/?p=689 [17:37] SketchCow: http://fromthepage.balboaparkonline.org/display/display_page?ol=w_rw_p_pl&page_id=1363#page/n0/mode/1up [19:03] Nice [19:09] Aww yeah [19:09] Got my new VDSL2 hooked up. [19:09] 263.90kB/s uploading to alard [19:09] alard: more anyhub is coming! [20:57] and i just installed the gem connection_pool [20:57] ok i got ruby gems to install finally, and their all setup, except im still getting this error ffgrab.rb:1:in `require': no such file to load -- connection_pool (LoadError) [21:01] bsmith093: ruby -v [21:02] actually, just send me your full terminal log [21:02] ruby 1.8.7 (2010-01-10 patchlevel 249) [i486-linux] [21:02] connection_pool does not work with Ruby 1.8.7, because it uses BasicObject, which only exists in Ruby 1.9 [21:02] also, Ruby 1.9 automatically loads Rubygems; 1.8.7 doesn't [21:02] apt install ruby1.9 [21:02] which is where the error you're seeing comes from [21:02] cause i think i did that [21:03] ruby1.9 is already the newest version. [21:03] ruby1.9 -v [21:03] ruby 1.9.0 (2008-10-04 revision 19669) [i486-linux] [21:04] ugh [21:04] that's...way behind [21:04] ah, another repo? [21:04] Ruby (and projects like it) move too fast for Debian/Ubuntu to keep up, IMO [21:04] oh wait yeah i just noticed the 2008 thing, wow, thats old [21:04] unless I can control the Ruby packages (e.g. for production environments) I use https://rvm.beginrescueend.com/ [21:05] it bypasses package management, but for me, the benefit outweighs that cost [21:07] got rvm now, grabbing ruby 1.9.3 [21:07] should i dump the ubuntu repo ruby? [21:07] only if you want to, it's not necessary [21:07] to dump it [21:08] k then, will this install it like a normal package? [21:08] RVM does not use apt, so no [21:08] lol @ a language moving so fast you can't package it [21:09] it will, however, modify your environment's PATH to work out [21:09] ersi: it's not that uncommon [21:09] yeah ive never heard of that [21:09] sounds more like a dialect, that forks all the time [21:09] I actually more often construct development environments directly from upstream than I do via OS packages [21:09] although i must say, this is the smoothest, complex thing i ve ever done [21:10] how do i keep it updated? [21:10] ersi: in particular, I've found that following upstream directly pays off for Node.js, factor, and GHC [21:10] bsmith093: rvm install [Ruby version] [21:11] so i have to know the version, i have , or the version i want to get? [21:11] ersi: also, the syntax and semantics of Ruby don't change that often (although ruby-core has been doing some WTFs in that regard lately) [21:11] ersi: the libraries, on the other hand [21:11] bsmith093: yes; rvm list will show you those [21:11] oh, wow this is cool, ive also never had this much feedback from a compiler that i could actually follow [21:12] I'm having a hard time understanding how a 10+year language can move so fast it's bleeding edge all the time [21:12] the language itself does not [21:12] implementations and libraries do [21:14] you know what would be nice? a dummy package for every linux distro, that does [language]-all, and grabs everything in the repos for that language [21:14] that would be infeasibly huge [21:14] how big could that possibly be? [21:14] for Ruby alone there's 31,503 libraries [21:15] Java would be an order of magnitude larger [21:15] mother of Turing, that's a lot of development [21:15] and to be fair, java mostly take care of it self as it needs to [21:16] keep jvm updated and afaik thats all u need to worry about [21:16] Hackage lists, uh [21:17] something around 3633 packages for Haskell [21:17] ok, ok, so languages are much bigger that I thought, in their entirety [21:18] yeah -- I find that a language is really nothing without its libraries [21:18] I mean, sure, you can install an implementation of a language [21:18] but it's really pretty useless on its own [21:18] hey another thing, does a sudo operation keep root until its done, or is their a timer somewhere? [21:19] because ive had things crap out asking for rights halfway through [21:19] rubys's done [21:20] annnnd.. same error as last time only this time ruby -v ruby 1.8.7 (2010-01-10 patchlevel 249) [i486-linux] [21:20] rvm use 1.9.3 [21:21] /home/ben/.rvm/rubies/ruby-1.9.3-p0/lib/ruby/site_ruby/1.9.1/rubygems/custom_require.rb:36:in `require': cannot load such file -- connection_pool (LoadError) [21:21] from /home/ben/.rvm/rubies/ruby-1.9.3-p0/lib/ruby/site_ruby/1.9.1/rubygems/custom_require.rb:36:in `require' [21:21] from ffgrab.rb:1:in `
' [21:21] paste me the full terminal outpuyt [21:22] /home/ben/.rvm/rubies/ruby-1.9.3-p0/lib/ruby/site_ruby/1.9.1/rubygems/custom_require.rb:36:in `require': cannot load such file -- connection_pool (LoadError) [21:22] ben@ben-laptop:~/1432483$ [21:22] from /home/ben/.rvm/rubies/ruby-1.9.3-p0/lib/ruby/site_ruby/1.9.1/rubygems/custom_require.rb:36:in `require' [21:22] from ffgrab.rb:1:in `
' [21:22] ruby ffgrab.rb [21:22] thats what i get [21:22] gem install bundler; bundle install [21:23] the Gemfile in the gist repo is a dependency manifest [21:23] for Bundler [21:24] i nthought that was important, i kept trying ruby Gemfile on the offchance something would happen, this is not an intuitive lang to install [21:25] Fetching source index for http://rubygems.org/ [21:25] now that seems like i would need that for gems, becasue thats where i found connection_pool and girl_friday [21:25] Rubygems is a packaging mechanism; bundler's a tool for managing packages [21:25] they're related, but Rubygems is independent of Bundler [21:26] well its finding the deps, and indtalling them , so whoo. [21:26] holy crap its running [21:27] I'd like to again point out that it doesn't do anything to record its results [21:27] and apparently, its timed itself to 6 decimal places? [21:27] times what [21:27] timestamp goes out to seconds.###### [21:28] that's the default behavior of the Ruby logger library [21:28] but, yeah, there's no point in running that as-is for a long time [21:28] man, thats precise [21:28] because it doesn't yet actually do anything aside from spit results to the console [21:29] I'm not even sure if it handles pages correctly -- I *think* it does, but I haven't run it long enough to see how they get processed in the queue [21:30] i just had a though, do user profiles show the stories all on one page, regardless of how many there are, cause that might be a help. [21:30] possibly, but AFAIK there is no way to get a list of all users [21:31] other than doing my original idea, and yours is much faster and uses less resources over all, on both ends [21:32] I can tell you that my method results in a lot of duplicates [21:32] in particular, it doesn't yet account for the "last page" link in each story [21:32] that will have to be filtered out in the discovery logic [21:33] yeah i dont really have any thoughts for that [21:33] it's just more HTML scraping [21:33] although the chapter is just a number appended to the link [21:33] not hard, just needs to be done [21:33] the next and back buttons are javascript, i think [21:34] https://gist.github.com/705cd333e06178057dec [21:34] that's a list of 4,506 story links recovered by ffgrab [21:34] well [21:34] 4506 / 2 roughly [21:34] wait the number before the title, thats the last chapter? [21:34] that's a chapter indicator [21:35] so let ffgrab run till its done then grep for dupes and keeep the higest number [21:35] I'd rather fix it in the grabber [21:35] to ignore it, you'll have to change what stories_and_categories_of does at lines 12-13 [21:36] I'm not sure what the change is, as I haven't looked at ff.net's page structure close enough to make the discernment [21:36] its still faster that iterating through 10mil semi fake links [21:37] I am also suspicious of results like this: [21:37] I, [2011-12-07T15:33:31.381205 #75544] INFO -- : Found 0 categories, 0 stories from /book/My_Sweet_Audrina/ [21:37] in that case, there really are no entries that show up [21:37] but any 0/0 results make me suspicious that the script is missing something [21:38] i was right, it is js here