[07:44] Torrent is at 5% [07:44] soultcer: oh, ok then. [08:35] Importing tasks.. [08:37] \o/ [08:46] http://urlteam.terrywri.st/ hello world [08:48] Would've preferred tracker.urlte.am as the URL, but oh well [08:50] Poke SketchCow and he'll fix the DNS for you. [08:51] SketchCow: tracker.urlte.am to 37.59.60.160 [08:51] much love [09:08] Great! [09:09] Make sure to run the cleanup.py script every 15 minutes or so to free assigned tasks [09:10] I'm sure cron wants some attention on anarchive [09:26] GLaDOS: Create a directory calles files in the tracker directory and make it writable for the webserver [09:27] done [09:28] That's where the output files will go until you get them with the fetch_finished.py script [09:30] Is it worth it running fetch_finished.py on cron? [09:30] s/on/with/ [09:31] not necessary but you could run it like once a day [09:31] The db gets kind of slow when there are lots of tasks, so I usually ran it every couple of days, put the results into a tar file and then later reimported that tar file into the main db [09:33] So for importing tasks into the main db, what format do you suggest them to be in? (tar, in a file, etc.) [09:34] * omf_ scribbles furious notes :D [09:36] The same format that fetch_finished spits out [09:36] I just tarred them up for transport because the tracker db and main db were on different hosts [09:37] Ah [09:37] It'd be so simple to make a cron script to do this. [09:37] If I had the time. [09:38] Which I technically do. [12:34] soultcer: it's still adding tasks [12:34] how long does it usually take? [12:34] (i forgot to run it in a screen session) [12:37] (chain_multiple(tracker, "bitly", "0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ", 6, 600, 10) and sequence_from_to(tracker, "tinyurl", "0123456789abcdefghijklmnopqrstuvwxyz", "azzzzz", "bzzzzz", 600), btw) [12:41] Well from azzzzz to bzzzzz with 36 characters is 60466176 codes, divided by 600 codes per task is 100777 tasks, of which you have already created 86000 [12:41] So another 14000 remaining [12:42] So, soon done ;) [12:42] 01[13tinyback01] 15twrist pushed 1 new commit to 06master: 02https://github.com/ArchiveTeam/tinyback/commit/12d8610b6cbfb49ba4ce060834a14755e255ffcf [12:42] 13tinyback/06master 1412d8610 15Terry Wrist: Change tracker to http://tracker.terrywri.st/ [12:43] IGNORE URL IN COMMIT MESSAGE, IT'S CORRECT IN CODE [12:44] I see graphs! [12:45] So other shorteners to add are ur1.ca, ow.ly, snipurl.com, vbly.us, arseh.at, and is.gd [12:50] let me fire up my warrior and pull down a few tasks to check it is all good [12:52] GLaDOS, you need to update http://warriorhq.archiveteam.org/projects.json to point to the new urlteam leaderboard [12:53] I don't have access [12:53] Smiley: ^ [12:53] Whos machine is that anyhow? [12:53] SketchCow, I believe [12:53] warriorhq? chronomex [12:54] xmc: ^* [12:54] Smiley also has access [12:54] We'll see who's the fastest [12:55] Bonus points to whoever gives us access [12:55] I will add an entry to http://archiveteam.org/index.php?title=Tracker [12:58] 01[13tinyarchive01] 15twrist pushed 1 new commit to 06master: 02https://github.com/ArchiveTeam/tinyarchive/commit/f5515b833b1f6a1331736b25352b377b01ef68fd [12:58] 13tinyarchive/06master 14f5515b8 15Terry Wrist: That's http://urlteam.terrywri.st/ of course.... [12:59] sequence_from_to(tracker, "isgd", "0123456789abcdefghijklmnopqrstuvwxyz", "kZZZZ", "zZZZZ", 600) [12:59] Will that work? [13:10] I updated the People section on the tracker page [13:18] soultcer, pass that ops around [13:23] We need to setup another admin for the projects file so 3 people have access [13:23] As we talked about way back it would be best if all services had 3-4 people as admins to better the odds of an admin being around in irc [13:24] I should set up access to pad/paste then [13:24] The URLTeam tracker is just running under user urlteam on anarchive [13:24] I saw and updated the wiki page with yourself, Smiley and myself as admins [13:36] sup? [13:37] Smiley, fix up the projects.json file to point to the new urlteam leaderboard tracker.terrywri.st [13:37] fuck [13:37] I mean urlteam.terrywri.st [13:39] hmmm ok let me see [13:39] "name": "URLTeam", [13:39] "description": "Archiving various URL shorteners", [13:39] "logo": "http://archiveteam.org/images/9/9d/Urlteam-logo-reasonable-size.png", [13:39] "repository": "https://github.com/ArchiveTeam/tinyback.git", [13:39] "title": "URLTeam", [13:39] "marker_html": " - Ur1ca's highest is frv3a, not fzzzz [14:07] is this data from the wiki page? [14:08] no [14:08] he just checked. The 'highest' changes constantly [14:08] GLaDOS: Depends on how much work there is @ re 600 URLs per item [14:09] I've checked arsehat [14:09] arsehat as well, they are only up to 4hc5 [14:10] We should have a note on the wiki that highest urls need to be checked and how to do that. [14:10] And you seem to have confused visibli (vsb.li/shared.co) with vbly.us (formlery vb.ly) [14:11] ah [14:11] Oh and the new tr.im is also net yet up to 50zzzz [14:12] Besides bit.ly, tinyurl.com maybe the bit smaller ones like is.gd and ow.ly, the other shorteners don't get many new URLs [14:12] yeah, added an extra z [14:13] http://paste.archivingyoursh.it/wageluxutu.md [15:40] soultcer: what happens if I add a task that's already added? [21:06] GLaDOS: The task will be run a second time, and then on import into the main db it will hopefully have no conflicts because none of the data on the url shortener is supposed to change [21:06] But there is no "deduplication". Tasks get an uuid and are always refered to by that id instead of the contents