[03:52] It this really the most updated code for the warrior? https://github.com/ArchiveTeam/warrior-code [09:14] omf_: Yes, do check if there's a development branch though [09:59] omf_, It's warrior-code2, not warrior-code [10:05] got it [10:05] I have built some deployable linux vms before. I am also interested in looking at how people do it. [10:06] Well the actualy building of the vm happens in another git repo called warrior-preseed [10:06] It's basically just a customized debian install using the debian installer's preseeding options [10:06] I'm currently playing around with it because I want to slim it down a bit [10:10] yeah I am familiar with the preseed features. I used it to build custom ubuntu installs for a non-profit [10:10] It really sped up their process [10:11] One trick I use is to scrub stuff out on the end of build hook [10:13] There is a script in warrior-code2 that removes all unneeded files [10:13] And some aptitude purge magic that deletes unwanted packages in the preseed file itself [10:13] Have you done any benchmarking [10:13] I know for size you have [10:14] what about RAM usage and boot speed [10:14] I don't think anyone has done such benchmarks, but you would have to ask alard to be sure. He is the genius that came up with the warrior [10:15] kiwi and cobbler are pretty advanced in those kinds of things [10:16] I never tried these, but my next step after cleaning up the package lists is to write a plugin for this: https://github.com/andsens/ec2debian-build-ami [10:16] Then we could have a warrior AMI that anyone can just run [10:17] kiwi is the opensuse distro builder, cobbler is the redhat distro builder [10:17] they both build to multiple formats by default including EC2 instances [10:18] the opensuse ones can be built and deployed from just the web [10:18] Well we are a debian shop, so we have to use what debian gives us :D [10:20] You should use what is best and everything is open source. Hench benchmarks [10:21] Like deploying a multithreaded scraper for example has different bottlenecks [10:23] I think the problem is that we don't have the manpower to do all that [10:23] most of it can be automated away [10:23] so not too much people power [10:24] We just need a slow strech [10:24] not fucking everything is dying at once bullshit [10:26] You forget that Archiveteam is a hobby project. Even if there is no project dying right now, there is not always time for new features [10:27] I am talking about just my personal time there [10:27] Oh, well then code away ;-) [10:27] All the code we have is on github [10:28] I have been looking through it [10:28] One thing which I think is important though is that more than one member needs to understand each project. [10:29] yes I agree [10:29] the hit by a bus problem [10:29] all the code is shared but is all the knowledge and process documented [10:30] The warrior documentation is pretty good and we have many people who are Debian users who can help [10:31] On the other hand we also have projects where the code is pretty much a mess and only one person knows how everything works [10:31] do we have good references for warc and cdx files [10:31] like I had to read the iso standard [10:31] and shit like that sucks the life out of you [10:32] Well, warc is an ISO standard so that pretty much is the reference [10:33] The Internet Archive can only add warc files to the wayback machine, not tar files made with wget, so we are stuck with that [10:33] it can handle the compressed warcs right? [10:34] It's just gzip compression of the whole file I think [10:35] But I never had to work with warc files, so in doubt ask alard or underscor ;-) [10:45] Most my current work is on url mapping, domain mapping and other things to make sure content coverage is good [10:45] all this can then be folded into preventative backups of key sites [11:16] omf_: The most important part of the warrior is the ArchiveTeam/seesaw-kit repository. You should install that and use the run-warrior command to start the warrior. ArchiveTeam/warrior-preseed and ArchiveTeam/warrior-code2 are specific to the vm image. You could use these as inspiration for your own. (It's very useful to use the same Debian distribution, though, or you'll have to compile your own Wget+Lua binaries.) [11:17] omf_: Benchmarks really depend on the project. Wget doesn't need a lot of memory, unless it finds a site with a lot of URLs, then no amount of memory is enough. [11:18] The warrior VM uses 400MB of RAM, and that has generally worked so far. (And that's about the only benchmark we have.) [11:19] That is interesting to know [11:21] omf_: The ISO standard for WARC isn't that bad, I think, and there's also the http://www.netpreserve.org/resources/warc-implementation-guidelines-v1 to go with it. [11:22] It's probably important to add that "compressed warcs" are compressed *per warc record*, so you can easily extract individual records. It's not just a gzipped file. [11:24] I am going to read that pdf alard [13:07] alard: Is there a reason why you have build-essential installed on the warrior vm? [13:10] Not that I know of. I started with a normal Debian installation and removed the things I thought could be missed. [13:10] Is is very large? [13:11] Perhaps I thought that compiling things was necessary, for pip installs, and this was "essential". [13:11] Well it pulls in a gcc so it takes a little bit of space [13:11] I am trying to get a list of all "warrior required" packages so that I can create an amazon ec2 image [13:11] gcc is probably useful. [14:42] iirc, one of the dependencies of seesaw has a module that compiles a C or C++ accelerated version if you have the proper dev packages installed [14:43] soultcer [14:43] simplejson [14:46] Thanks [14:48] I think it needs build-essential and some python dev package [15:56] alard: btw, I pushed the rest of those changes to my pull request. want to merge them or should I just do it? [16:50] chazchaz: I just set up the warrior without gcc and you are absolutely right, simplejson complains that it will be installed without setups [16:50] But shouldn't the json module in python 2.6 work just as well/fast? [16:50] it works just as well, and I doubt the speed difference is noticable [16:51] all we need to do is parse the occasional request for more work [16:56] soultcer: No, each WARCRecord is gzipped. Not the whole file [16:56] So it's basicaly a bunch of gzip streams in one file [16:57] oh, I missed that alard filled that in already. Sorry. [23:19] Is this the best channel we have to talk about scrapers?