#newsgrabber 2017-08-03,Thu

Logs of this channel are not protected. You can protect them by a password.

↑back Search ←Prev date Next date→ Show only urls(Click on time to select a line by its url)


WhoWhatWhen
***criz2 has quit IRC (Ping timeout: 246 seconds)
criz2 has joined #newsgrabber
[00:52]
.............................................................................................................................................................................................. (idle for 15h47mn)
HCross has quit IRC (Read error: Connection reset by peer) [16:39]
HarryCros has joined #newsgrabber [16:45]
...................................... (idle for 3h9mn)
arkiverhmm
HCross2: discovery seems to be not working so well today
[19:54]
HCross2hang on - let me investigate
I think I know what may have happened
[19:55]
arkiverwhat do you think happened?
arkiver is busy setting laptop up for the next 5 days
need to have all to be able to work on stuff
[19:58]
HCross2arkiver: I had my bank card stolen a couple of weeks back.. all the servers that do discovery were billed to that card. Of course they were all declined this month [20:00]
arkiverouch...
that sucks
nothing stolen from your card after it was stolen?
with your card*
[20:02]
HCross2Nope. I use an app bank so I was able to suspend the card straight after
arkiver: you heard of Bunq?
[20:04]
arkivergood
yes, I've heard of them
not using them though
[20:04]
HCross2Ah. I'm with Starling which is sort of the British alternative
Both discovery are coming back
[20:05]
arkiverRight
I do know that bunq recently increased prices, which gave them quite some bad attention
awesome, thanks
it also seems like the number of discovered URLs recently is much lower than what we used to get
used to be around 7 or 8 thousands or so
now only around 4 to 5 thousand
arkiver is afk for 30 minutes
it's nice how there's always less websites shutting down during the summer vacation period
[20:05]
HCross2arkiver: one thing I noticed on my grabbers.. netstat shows an absolute ton of requests to Google.com [20:10]
arkiverHCross2: interesting
not sure what's up with that
afaik we're not archiving google that much
[20:15]
HCross2Check the Copenhagen one I sent yoy [20:15]
***n00b038 has joined #newsgrabber [20:24]
n00b038.
.
.
.
.
[20:24]
***n00b038 has quit IRC (Client Quit) [20:24]
arkiversorry that was me testing a little
HCross2: yes, will have a look at that
[20:25]
..... (idle for 24mn)
HCross2: are you running the project on the copenhagen server?
if not I'm going to clean it up a bit
there's a lot of tmp-* files for some reason
[20:49]
HCross2It is running a grabber, it loves making temp files for some reason
Feel free to stop and restart it if needs be
[20:50]
arkiverI don't have to password for su
can you stop the crawls in screen run by archiveteam?
[20:58]
well you don't have the wpull.db problem
it was just filled up with tmp files
[21:04]
I don't think it's possible to prevent all loos
loops
so it might be best to set a limit of for example 100000 URLs
what do you think?
[21:13]
.................... (idle for 1h38mn)
***HCross has joined #newsgrabber
HarryCros has quit IRC (Ping timeout: 268 seconds)
[22:51]

↑back Search ←Prev date Next date→ Show only urls(Click on time to select a line by its url)