#urlteam 2014-02-13,Thu

↑back Search

Time Nickname Message
00:20 🔗 chfoo the tracker is under GLaDOS domain though
00:23 🔗 GLaDOS i would be maintaining it but year 11 happened
00:46 🔗 chfoo i have a crazy idea. create a github organization with lots of repos for the shorteners. the scraper scripts will scrape and upload to a staging server which will then commit the urls to github. this way, it allows easy browsing urls and selective downloading of the data set.
02:49 🔗 GLaDOS so a whole restructure?
02:50 🔗 GLaDOS could work
04:45 🔗 xmc chfoo: the data is *way* too big for github
04:50 🔗 chfoo maybe github won't notice if we have hundreds of repos :p
05:17 🔗 xmc we're all going to internet jail
06:31 🔗 GLaDOS ono
06:40 🔗 SketchCow Don't drop the internet soap
11:41 🔗 ersi chfoo: GitHub allows only 1GB of data in each git repo.
13:22 🔗 GLaDOS LET'S UPLOAD IT TO IA THEN!
14:14 🔗 ersi Well, yeah - that'd be the obvious good candidate storage place
14:14 🔗 ersi But if I understand correctly, the problem isn't hosting for the data set
14:49 🔗 GLaDOS ..what was it?

irclogger-viewer