[00:01] godane: --span-hosts -D hackadaycom.files.wordpress.com,hackaday.com [00:01] they do that as load balancing [00:02] serving static images from another domain = another server plus browsers can handle it better [00:03] i been downloading hackaday for 2 hours [00:03] so i'm not going to do that now [00:05] i'm getting the images thur a script i wrote [01:19] maybe Jason Scott should get this: http://www.ebay.com/itm/MTV-Presents-Spyder-Games-Soap-Opera-VHS-Video-Complete-Series-Sunset-Beach-OOP-/290705456005?pt=VHS&hash=item43af653b85 [01:42] how do you stop wget-warc from downloading anything with "?replytocom=" and "?shared=" [02:20] just got the youtube embed list [02:21] from hackaday.com [02:23] fucking A [02:24] got youtube list to not have ' in from http [02:24] *front [02:36] good news everyone [02:37] i can maybe download all youtube embeds videos off of hackaday.com [04:30] Do the A/V Geeks have a shirt design up? I'd get the $50 if it's a nice shirt [08:58] are SketchCow and underscor at HOPE or something? [08:58] i think SketchCow is at defcon [08:59] or that [08:59] anyways, I changed my attack on archiving hackaday.com [09:00] i do hackaday.com/2004/12/ and just get all hackaday data from dec 2004 [09:02] this way it shouldn't take forever [09:02] i downloads a lot of ?shared=service crap and i don't kow how to make that stop [09:04] --reject-regex= maybe [09:05] AT really could use a blog or something [09:05] or tutorials, howtos, howwedid [09:07] there is --reject option [09:07] doesn't work for me [09:08] and when i reject does work it will still download it [09:08] it just will not store it [09:08] yeah, that one is weird. i think the regexp one works with files like that [09:08] oh :( [09:10] i get stuff like index.html?share=*.html crap [09:10] i hate it now [09:11] its meant to share stuff on facebook and digg and email [09:54] i still can't get reject to work that way i want it too [09:54] Schbirid: can you help me with this? [09:55] i want wget to only download index.html files [09:55] i could try with no garantuee (how do you spell that...) [09:55] wget-warc "http://$1/2004/12/" --mirror --warc-file="$1-2004-12" --warc-cdx -np -A "index.html" -R "*\?" --html-extension --domains hackaday.com -p [09:55] thats my code for my script [09:56] --reject-regex='/.*?' maybe? [09:57] example url will be stuff like this: http://hackaday.com/2004/12/31/treo-650-hack-how-tos/?shared=email&msg=fail [09:58] --reject-regex doesn't work [09:58] it doesn't even exist [09:58] oh yuck, i just realised. i am using a very recent wget and thatoption is not even on the manpage :( [09:59] that works well though [09:59] wget -m -np --reject-regex='/.*' "http://hackaday.com/2004/12/31/treo-650-hack-how-tos/" [09:59] only gets the root (index.html) for me [09:59] GNU Wget 1.13.4-2608 [09:59] oh [09:59] i'm using 2582 [10:01] hm, not sure how to make it greedy though. wget -m -np --reject-regex='/.*?' "http://hackaday.com/2004/12/31/" does not traverse into the "subdirs" [10:01] lunch time [10:10] maybe --reject-regex='/.*\?.*' [10:10] as in reject anything with a ? [10:11] oh, '\?' would work then :D [10:11] 2012-07-28 12:10:02 URL:http://hackaday.com/2004/12/31/ [43557] -> "hackaday.com/2004/12/31/index.html" [1] [10:11] 2012-07-28 12:10:03 URL:http://hackaday.com/2004/12/31/treo-650-hack-how-tos/ [74383] -> "hackaday.com/2004/12/31/treo-650-hack-how-tos/index.html" [1] [10:11] 2012-07-28 12:10:04 URL:http://hackaday.com/2004/12/31/treo-650-hack-how-tos/feed/ [15930] -> "hackaday.com/2004/12/31/treo-650-hack-how-tos/feed/index.html" [1] [10:11] 2012-07-28 12:10:05 URL:http://hackaday.com/2004/12/31/treo-650-hack-how-tos/ [74383] -> "hackaday.com/2004/12/31/treo-650-hack-how-tos/comment-page-1/index.html" [1] [10:12] looks good [10:14] cool [10:14] thanks [10:16] np [10:29] looks like wget-warc can't be complied without downloading something [10:29] :-D [10:29] :-( [10:29] the last one i meant [10:30] i'm been trying to make a source dvd that can recompile a os fully offline [11:31] Schbind: i'm starting to hate wget-warc [11:31] i can't compile it like the normal tarballs [11:33] godane: i am using a bazaar checkout.wget has warc now [11:34] i hate it cause i want to compile it offline [11:35] no download crap in order to compile [11:35] i don't want gnulib being some depend for it [12:03] i got error even when doing it right [12:03] damn wget-warc [12:08] i can't build the newer wget-warc [12:09] Schbirld: can you help me? [12:10] there is no css.c file wget [12:10] *in wget source [12:10] thats my error [12:33] looks like i maybe able to just pack the old wget-warc with regex support [12:33] nothing got broken out side of changelogs [12:56] its working now [12:56] only download stuff thats need now [14:40] no idea, sorry [19:09] http://vimeo.com/43530099 [21:52] it will sound bonkers to you winr4r as it sounds bonkers to me evne now [21:52] but I overcame my depression by willing myself better :< [21:54] SmileyG: i solve it by being insanely productive [21:55] SmileyG: among many other things i'm doing today, i'm writing a piece explaining all the cultural and historical references in "alice's restaurant" [21:57] SmileyG: i think that a lot of kids in the west wouldn't get most of the jokes *today*, let alone someone from another culture who heard it [22:00] I have no clue what you just said. [22:00] sorry to hear that [22:01] What is Alice's Resturant? [22:01] err spelling :D [22:02] SmileyG: http://www.youtube.com/watch?v=LjKF7aQthcQ [22:02] it is actually a 20-minute song btw [22:02] which is another reason it is so funny, most of arlo's songs were of normal length [22:03] this reminds me of tribute [22:03] This isn't the best song in the world, its merely a tribute... [22:07] good song too [22:07] * SmileyG still listening... [22:11] SmileyG: it's a later version of the song, played live, so there's quite a few self-references in there [22:14] * SmileyG doesn't understand [22:14] you'll see towards the end [22:20] o_O [22:23] the thing about 18 minutes and 20 seconds, by the way, might justify having a whole new appendix for alternative endings [22:25] hmmm [22:25] Ok, I didn't understand any of that :/ [22:26] SmileyG: didn't understand the accent, or understand the cultural references? [22:27] I didn't even notice any cultural references [22:28] there are lots of them [22:28] 8x10 pictures? [22:28] the significance of that is quite important [22:28] ...? [22:29] * SmileyG is at a loss. [22:29] 8x10 film cameras are what you shot back then when you wanted ultimate technical quality [22:29] Right... [22:29] that would be like the police photographing littering with £20,000 hasselblads [22:30] hmmm ok? [22:30] XD [22:30] * SmileyG never understood people reading into songs. :/ [22:30] Paintings yes, I can get that [22:30] Songs.... nope, they are "just songs" to me :S [22:30] Then again some of the stuff people read into paintings I think is pure BS too :/ [22:31] SmileyG: i'm not reading into it, it's how people would have understood the song 45 years ago [22:31] artistic licence because it rhymes? [22:33] nothing in that song rhymes :P [22:33] you know what I mean :P [22:33] * SmileyG just doesn't understand [22:34] :S [22:34] "8x10 colour glossy photographs" isn't because it rhymes, it's because "8x10 colour glossy photographs" had significance back then [22:34] I've tried to explain to people how i experience the world before and I fail to [22:34] 1) it was 8x10 which was overkill for photographing a crime scene [22:34] or they fail to comprehend what I experience. [22:34] 2) it was in colour, at a time when colour processing was expensive [22:34] which the cops used to photograph littering in a very very small town [22:35] So the song is deliberately absurd? [22:35] SmileyG: yes, but apparently a true story [22:36] SmileyG: you only get the full feel of the absurdity once you understand all the historical/cultural references [22:36] winr4r: now I understand [22:36] to a present day listener its just a song [22:37] but when you know the cirucmstance you understand how absurd it is. [22:37] yup, which is the whole reason i am writing [22:45] hi mistym [22:45] Hey winr4r [23:08] hey winr4r [23:08] i'm backing up hackaday.com [23:08] godane: yes, i saw that :) [23:08] i doing it by year [23:09] i have 2004 and 2005 backed up [23:09] also got data dumps of images [23:10] also uploading gbtv episode 2012-01-16 [23:10] you the man :) [23:11] learned about grep -o option [23:12] makes alot easier to grep all www.weblogsinc.com/common/images/[0-9]*.JPG