[01:01] *** icedice has quit IRC (Quit: Leaving) [01:42] *** dashcloud has joined #archiveteam-ot [02:11] *** ivan` has joined #archiveteam-ot [02:11] *** vectr0n_ has joined #archiveteam-ot [02:11] *** djsundog has quit IRC (Ping timeout (120 seconds)) [02:11] *** vectr0n has quit IRC (Excess Flood) [02:11] *** dxrt_ has quit IRC (Ping timeout (120 seconds)) [02:11] *** Albardin has quit IRC (Write error: Broken pipe) [02:12] *** Odd0002 has quit IRC (ZNC - http://znc.in) [02:12] *** ivan has quit IRC (Remote host closed the connection) [02:12] *** faolingfa has quit IRC (Remote host closed the connection) [02:13] *** vectr0n_ is now known as vectr0n [02:16] *** nightpoo- has quit IRC (ircd.choopa.net irc.choopa.net) [02:16] *** Polylith has quit IRC (ircd.choopa.net irc.choopa.net) [02:16] *** moufu has quit IRC (ircd.choopa.net irc.choopa.net) [02:16] *** chfoo has quit IRC (ircd.choopa.net irc.choopa.net) [02:16] *** arkiver has quit IRC (ircd.choopa.net irc.choopa.net) [02:16] *** jut_ has quit IRC (ircd.choopa.net irc.choopa.net) [02:16] *** kiska1 has quit IRC (ircd.choopa.net irc.choopa.net) [02:16] *** mal has quit IRC (ircd.choopa.net irc.choopa.net) [02:16] *** tyzoid has quit IRC (ircd.choopa.net irc.choopa.net) [02:16] *** sep332 has quit IRC (ircd.choopa.net irc.choopa.net) [02:19] *** faolingfa has joined #archiveteam-ot [02:22] *** moufu_ has joined #archiveteam-ot [02:24] *** chfoo-_ has joined #archiveteam-ot [02:26] *** Odd0002 has joined #archiveteam-ot [02:26] *** nightpoo- has joined #archiveteam-ot [02:26] *** Polylith has joined #archiveteam-ot [02:26] *** arkiver has joined #archiveteam-ot [02:26] *** jut_ has joined #archiveteam-ot [02:26] *** kiska1 has joined #archiveteam-ot [02:26] *** tyzoid has joined #archiveteam-ot [02:26] *** sep332 has joined #archiveteam-ot [02:26] *** djsundog has joined #archiveteam-ot [02:28] *** kiska1 has quit IRC (Read error: Operation timed out) [02:28] *** nightpoo- has quit IRC (Read error: Operation timed out) [02:28] *** tyzoid has quit IRC (Write error: Broken pipe) [02:28] *** nightpool has joined #archiveteam-ot [02:28] *** arkiver has quit IRC (Read error: Operation timed out) [02:29] *** jut_ has quit IRC (Read error: Operation timed out) [02:29] *** jut_ has joined #archiveteam-ot [02:29] *** sep332 has quit IRC (Read error: Operation timed out) [02:30] *** Polylith has quit IRC (Read error: Operation timed out) [02:30] *** djsundog has quit IRC (Read error: Operation timed out) [02:30] *** Polylith has joined #archiveteam-ot [02:32] *** arkiver has joined #archiveteam-ot [02:41] *** mal has joined #archiveteam-ot [02:42] *** kiska1 has joined #archiveteam-ot [03:27] *** Albardin has joined #archiveteam-ot [03:27] *** tyzoid has joined #archiveteam-ot [03:27] *** dxrt_ has joined #archiveteam-ot [03:27] *** dxrt sets mode: +o dxrt_ [03:27] *** djsundog has joined #archiveteam-ot [03:33] *** sep332 has joined #archiveteam-ot [03:48] *** odemg has quit IRC (Read error: Operation timed out) [04:00] *** odemg has joined #archiveteam-ot [04:10] *** DrasticAc has joined #archiveteam-ot [04:46] can't remember if i asked here before or not, but is there a preferred way to download entire mediawiki wikis? (entire meaning images, page edits, audio clips, files, etc.) [04:47] for example, i'd like to get an offline copy of https://wiki.telefang.net/ [04:47] btw i not only want to download it but be able to view it too [04:55] can anyone reccomend an iso burning software for macosx Kiska or w0rmhole ? [04:55] finder can do that already [04:56] pop in your disc, right click the disk image, and click burn disk image to disc" [04:56] right click, or control+click, or two-finger click [04:56] I want to make it an ISO from a disk [04:57] open disk utility [04:57] highlight the disc in the list [04:57] go to file>new image>image from [04:58] Ok cause I have a game that is not the IA as far as I know [04:58] be sure to select format as "dvd/cd master" [04:58] what's the game? [04:59] https://www.ebay.com/itm/Mega-Bloks-Dragons-Metal-Ages-PC-Game-Program-Windows-Mint-/173527845467?oid=172724424844 [04:59] only mention I could find of it [04:59] my assumption is that it has no copy protection so doing it the disk utility way should work [05:00] I come across a lot of obscure stuff [05:01] then it will give you a .cdr file. just type `hdiutil convert /path/to/file.cdr -format UDTO -o /path/to/converted.iso' [05:01] https://archive.org/details/@flashfire42 [05:01] cool :D [05:01] I just convert it to an ISO in the about this file thing is that bad? [05:01] like rename .cdr to .iso? [05:01] yeah [05:01] its worked in the past [05:02] I think [05:02] i though i read somewhere that that works, but i've looked and can't find that [05:02] *** Albardin has quit IRC (Read error: Operation timed out) [05:02] *** sep332 has quit IRC (Write error: Broken pipe) [05:02] *** tyzoid has quit IRC (Write error: Broken pipe) [05:02] *** kiska1 has quit IRC (Read error: Operation timed out) [05:02] *** djsundog has quit IRC (Read error: Operation timed out) [05:02] *** dxrt_ has quit IRC (Read error: Operation timed out) [05:02] netshit [05:02] Well if you want to try some of my old games go ahead [05:02] i mean netsplit [05:02] :) [05:02] The ones I have uploaded to the archive already [05:02] i'd love to [05:03] they are mostly going to need to be run through WINE or crossover if you dont have a windows device handy [05:04] there is a proper way to create as-close-to-original-as-possible copies of discs, but it requires expensive and specific hardware [05:04] Yeah I dont have that kind of hardware though if someone does I would be happy to send them the disks [05:05] godane might he does a lot of digitisation and shit [05:05] but w0rmhole if you want to try out some of the old programs I have uploaded be my guest [05:06] i have the hardware (a winxp box, ide plextor px-708a) but not enough time, and i need to reinstall xp on there anyway [05:06] sure i'll check them out [05:06] http://wiki.redump.org/index.php?title=CD_Dumping_Guide_(DIC_CLI) [05:07] has a list of archive-grade drives [05:08] my px-708a was about $165 AUD and i had to go through three to finally get a working one [05:08] was a lot of money to dump just like that but, archiving is what i live for [05:09] btw flashfire if you have a scanner i would really appreciate it if you could scan your discs too [05:10] the disc itself or whatever the jewel case artwork is if you have it [05:10] just to make as complete of a copy as possible [05:10] I dont have a scanner [05:10] Sadly [05:11] could you take a photo of the discs? [05:11] Though if you have the equipment and would like me to keep an eye out for sales just let me know [05:11] yeah [05:11] I can and upload that [05:11] I also need to have another look I want to buy some rom dumping equipment [05:12] I have some bootleg GBA games somewhere around [05:12] for something like that one where its relatively inexpensive and in great condition i could buy that [05:12] if you need suggestions on archival hardware feel free to ask me anytime [05:13] *** m007a83 has quit IRC (Read error: Operation timed out) [05:13] for dumping gba roms, you could always buy a nintendo ds flashcart (like an R4) and put a gba rom dumping utility on there. ive done that before [05:13] with a ds lite [05:13] cant remember the exact name of the tool though [05:13] *** m007a83 has joined #archiveteam-ot [05:14] but i did check the sha1s and md5s of the retail cartridges i dumped and they matched their no-scene counterparts [05:15] wait you do GBA rom dumping? [05:15] Do you have the files? [05:15] when i have carts worth dumping, yes [05:15] which files? [05:15] If you can get me the files I can upload the GBA rom files to the internet archive for you plus add them to my personal collection [05:16] sure, but which files? the roms i dumped? the actual dumping software? the r4 firmware? [05:17] i too have an IA account btw, but i haven't added that much stuff in a while [05:17] and i don't upload my tubeup videos to it [05:17] *** BlueMax has quit IRC (Read error: Connection reset by peer) [05:17] The roms you dumped [05:17] I would be very interested in them [05:18] also if you havent already check out http://hhug.me/ [05:22] most of them were retail by the way, three of them were firered rom hacks. and yeah i've been there before, great site. [05:22] it's been years since the last time i dumped those, i'd have to scour my hdds to find them [05:22] PLEASE DO SO [05:22] I dont mind whatever they were [05:23] Even a slightly different hash is good for me [05:32] ok hold your horses while i go look [05:35] i think that's the fastest i've ever found something [05:36] lol [05:36] where will you upload them? [05:36] dmca.gripe [05:36] dmca.gripe mangles the filename though, so i'll store it in a .7z [05:36] *store them all in a .7z [05:37] alright upload them and dm them to me My laptop will go dead soon so Will have a look tommorow or later today probably [05:37] and by mangles i mean, if you upload `horses.png' it will turn out as `a2dfscxCGcx.png' [05:37] ok, should be just about five minutes or so [05:39] archiving them rn [05:39] *** kiska1 has joined #archiveteam-ot [05:41] there will only be one POKEMON_FIRE_BPRE01.gba file. two of the three rom hacks had the same name, as well as my official firered cart [05:42] so im not sure if the one in there is a hack or the retail [05:53] *** chfoo-_ is now known as chfoo [06:01] *** Albardin has joined #archiveteam-ot [06:01] *** tyzoid has joined #archiveteam-ot [06:01] *** dxrt_ has joined #archiveteam-ot [06:01] *** dxrt sets mode: +o dxrt_ [06:01] *** djsundog has joined #archiveteam-ot [06:05] *** sep332 has joined #archiveteam-ot [06:32] *** m007a83_ has joined #archiveteam-ot [06:34] *** Meroje has joined #archiveteam-ot [06:35] *** m007a83 has quit IRC (Ping timeout: 252 seconds) [06:35] *** m007a83_ is now known as m007a83 [06:59] *** BlueMax has joined #archiveteam-ot [07:15] *** sep332 has quit IRC (hub.efnet.us ircd.choopa.net) [07:15] *** djsundog has quit IRC (hub.efnet.us ircd.choopa.net) [07:15] *** dxrt_ has quit IRC (hub.efnet.us ircd.choopa.net) [07:15] *** tyzoid has quit IRC (hub.efnet.us ircd.choopa.net) [07:15] *** Albardin has quit IRC (hub.efnet.us ircd.choopa.net) [07:15] *** kiska1 has quit IRC (hub.efnet.us ircd.choopa.net) [07:15] *** mal has quit IRC (hub.efnet.us ircd.choopa.net) [07:15] *** jut_ has quit IRC (hub.efnet.us ircd.choopa.net) [07:15] *** moufu_ has quit IRC (hub.efnet.us ircd.choopa.net) [07:15] *** Sanky has quit IRC (hub.efnet.us ircd.choopa.net) [07:15] *** svchfoo3 has quit IRC (hub.efnet.us ircd.choopa.net) [07:15] *** rektide has quit IRC (hub.efnet.us ircd.choopa.net) [07:16] *** jut_ has joined #archiveteam-ot [07:16] *** moufu_ has joined #archiveteam-ot [07:16] *** Sanky has joined #archiveteam-ot [07:16] *** svchfoo3 has joined #archiveteam-ot [07:16] *** rektide has joined #archiveteam-ot [07:16] *** irc.mzima.net sets mode: +o svchfoo3 [07:23] is there a preferred way to download entire mediawiki or fandom wikia wikis? (entire meaning images, page edits, audio clips, files, etc.) [07:23] i believe that just throwing it into grab-site can run into stacking urIs [07:24] or something with login page issues [09:40] So dxrt what brought you to the archiveteam [09:42] And you kiska what brought you here? [09:42] Hmm probably reading about the geocities stuff and finding the wiki a good resource a good enough reason to join [09:42] Fair enough was curious is all [09:42] You? [09:43] I honestly dont remember I was looking into datahoarding and the internet archive and found this place mentioned somewhere [09:44] It is mentioned a lot on that subreddit lol [09:45] I had a habit of clicking on the "About this Capture" thing on WBM and that is how I found AT [09:45] How long have you been with the team kiska I seem to remember you joining after me [09:46] I joined IRC after you, I ran the warrior since version 2 [09:46] yeah [10:34] *** wp494 has quit IRC (Ping timeout: 492 seconds) [10:35] *** wp494 has joined #archiveteam-ot [10:42] *** BlueMax has quit IRC (Read error: Connection reset by peer) [11:26] #archivebot: "<+w0rmhole> why does that say "dick" in it?" -- On a related note, the URL I got when I archived Trump's tweets a while ago was https://transfer.sh/GayO1/realDonaldTrump-tweets :-) [11:30] xD [11:31] *** kiska1 has joined #archiveteam-ot [12:02] no flounce no fuss no drama just liek movin' on izzle [12:02] i dont really game tbh [12:02] dont reall has mich in uuuh... common with majority o' peeps here 'cpz of taht an' etcetc [12:02] i mena mah irl frienfs they game liek welp 0% so i feel iddi MUCH bettah than 'em! for abit [12:02] ima sho' you get it tho [12:02] tried to get snowy back failed too bad so sad much hashtag [12:02] Kiska JAA try and make sense of that [12:02] ... [12:02] No thanks [12:03] That’s a message from a friend of mine. Her pseudonym is Fairfox no one ever knows what the duck she is saying [12:07] That's disgusting. [12:09] How I summed it up or how she types [12:09] ? [12:09] The latter. [12:10] *** dxrt_ has joined #archiveteam-ot [12:10] *** dxrt sets mode: +o dxrt_ [15:51] *** mgrytbak has joined #archiveteam-ot [16:07] *** Mateon1 has quit IRC (Ping timeout: 255 seconds) [16:07] *** Mateon1 has joined #archiveteam-ot [16:22] 00:23 is there a preferred way to download entire mediawiki or fandom wikia wikis? || wikiteam has tooling exactly for this [17:06] *** godane has quit IRC (Ping timeout: 260 seconds) [17:11] *** moufu_ is now known as moufu [17:22] *** godane has joined #archiveteam-ot [17:23] *** svchfoo3 sets mode: +o godane [17:26] https://www.archiveteam.org/index.php?title=Deaths_in_2018 how does the bot on this page know when a website is saved? [17:26] archivebot just finished www.joefrank.com [17:29] is there a bot? i thought it was manual [17:34] I think it searches https://archive.fart.website/archivebot/viewer/ [17:35] near the top of the page it says not to edit it, as it is "automatically updated by bot". [17:35] ah ok [17:37] HadeanEon is the botname [17:55] VoynichCr is the author. [17:55] on the github page it says that using `--xml' and `--images' downloads "complete XML histories and images". that means it downloads images, page edits, audio clips, files, etc., right? [17:55] w0rmhole: I think it checks against the ArchiveBot viewer periodically to update the saved/not saved thing. [17:55] im just confused on the terminology [17:56] *** vectr0n_ has joined #archiveteam-ot [17:57] *** vectr0n has quit IRC (Ping timeout: 259 seconds) [17:57] *** vectr0n_ is now known as vectr0n [17:58] JAA: i see [18:13] w0rmhole: bot updates page once a day [18:13] a more frequent rate would flood page history and recent changes [19:08] *** wp494 has quit IRC (Quit: LOUD UNNECESSARY QUIT MESSAGES) [19:09] *** wp494 has joined #archiveteam-ot [19:27] thanks voynichcr and jaa and everyone else [19:39] kiska, flashfire: macOS Mojave comes out today!!! :D [19:41] *** caff has joined #archiveteam-ot [19:53] JAA: does snscrape grab the urls from retweets? [19:54] *** SimpBrain has joined #archiveteam-ot [19:57] w0rmhole: Nope. [19:57] do you have plans of it grabbing those urls in the future? [19:58] If you know a way to discover them, sure. [19:58] i only know bash [19:58] Note that the profile page is limited to 3200 results. That's why snscrape uses the Twitter search, not the profile page. [19:58] oh i didnt know that [19:58] And there doesn't seem to be a way to find all tweets retweeted by @user through the search. [19:59] damn [20:30] so how many tweets can find in Tw search? your best results JAA ? [20:37] VoynichCr: All of them, but only original tweets by the user. The largest accounts I've scraped so far had roughly 150k tweets. [20:48] JAA: thats pretty cool, is snscrape going to be added to archivebot soon? [20:52] *** S1mpbrain has joined #archiveteam-ot [20:52] VoynichCr: "Soon"? Not likely. [20:53] *** S1mpbrain has quit IRC (Remote host closed the connection) [20:54] :-( [20:54] ArchiveBot development is moving at glacial speeds. [20:55] *** SimpBrain has quit IRC (Read error: Operation timed out) [20:57] not bad, it could be slower [20:57] Heh [20:57] Yeah, it could be 0 commits per year instead of 1. [21:03] one would think Commons and Wikipedia pages are well archived, but this file (and I have found many others) from year 2015 is yet to be archived https://web.archive.org/web/*/https://commons.wikimedia.org/wiki/File:Bulletin_of_the_Museum_of_Comparative_Zoology_at_Harvard_College_(1982-1985)_(20493697735).jpg [21:03] (i know about the xml and image dumps, I mean the wayback) [21:05] 2011 https://web.archive.org/web/*/https://commons.wikimedia.org/wiki/File:West_Allen_Dale_-_Low_Hartleycleugh_and_High_Turney_Shield_-_geograph.org.uk_-_1703591.jpg [21:06] the wayback coverage of Commons sucks [21:15] I'd guess that it's intentionally excluded from IA's crawls because there's no risk of it suddenly disappearing anytime soon. [21:16] And IA is directly collaborating with Wikimedia in other aspects (e.g. prevet linkrot among sources on Wikipedia articles), so if Wikipedia got in trouble, they'd likely be able to archive it then before anything is lost. [21:20] there must be an explanation for sure, but it is a bit weird to me [21:21] one of the most popular websites in the interwebs, the biggest encyclopedia, free licensed, is poorly archived in the wayback [21:22] next year the Commons interface is gonna change, Structured Commons is comming soon [23:14] *** Odd0002 has quit IRC (Quit: ZNC - http://znc.in) [23:14] *** Odd0002 has joined #archiveteam-ot [23:16] *** caff_ has joined #archiveteam-ot [23:19] *** caff has quit IRC (Read error: Operation timed out)