[00:51] *** robogoat has quit IRC (Ping timeout: 258 seconds) [01:09] *** robogoat has joined #archiveteam-ot [01:18] *** yawkat has quit IRC (Ping timeout: 496 seconds) [01:31] *** yawkat has joined #archiveteam-ot [01:38] *** Ivy has joined #archiveteam-ot [03:08] *** manjaro-u has quit IRC (Read error: Operation timed out) [04:29] *** synm0nger has joined #archiveteam-ot [04:30] *** SynMonger has quit IRC (Ping timeout: 246 seconds) [04:36] *** qw3rty has joined #archiveteam-ot [04:43] *** qw3rty2 has quit IRC (Ping timeout: 745 seconds) [04:53] *** icedice2 has joined #archiveteam-ot [04:53] *** yano_ has joined #archiveteam-ot [04:55] *** voker57_ has joined #archiveteam-ot [04:55] *** benjinsmi has joined #archiveteam-ot [04:56] *** girst_ has joined #archiveteam-ot [04:57] *** Maylay_ has joined #archiveteam-ot [04:57] *** Maylay_ has quit IRC (Remote host closed the connection!) [04:58] *** Maylay_ has joined #archiveteam-ot [04:58] *** tuluu_ has joined #archiveteam-ot [05:00] *** chfoo_ has joined #archiveteam-ot [05:00] *** Fusl__ sets mode: +o chfoo_ [05:00] *** Fusl_ sets mode: +o chfoo_ [05:00] *** Fusl sets mode: +o chfoo_ [05:05] *** ibachandl has quit IRC (Quit: Page closed) [05:12] *** tuluu has quit IRC (irc.efnet.nl efnet.deic.eu) [05:12] *** icedice has quit IRC (irc.efnet.nl efnet.deic.eu) [05:12] *** dashcloud has quit IRC (irc.efnet.nl efnet.deic.eu) [05:12] *** benjins has quit IRC (irc.efnet.nl efnet.deic.eu) [05:12] *** girst has quit IRC (irc.efnet.nl efnet.deic.eu) [05:12] *** Maylay has quit IRC (irc.efnet.nl efnet.deic.eu) [05:12] *** voker57 has quit IRC (irc.efnet.nl efnet.deic.eu) [05:12] *** nepeat has quit IRC (irc.efnet.nl efnet.deic.eu) [05:12] *** fuzzy8021 has quit IRC (irc.efnet.nl efnet.deic.eu) [05:12] *** chfoo has quit IRC (irc.efnet.nl efnet.deic.eu) [05:12] *** wp494 has quit IRC (irc.efnet.nl efnet.deic.eu) [05:12] *** yano has quit IRC (irc.efnet.nl efnet.deic.eu) [05:18] *** IAmbience has quit IRC (Quit: Connection closed for inactivity) [05:29] *** fuzzy8021 has joined #archiveteam-ot [06:04] *** HP_Archiv has joined #archiveteam-ot [06:04] So furthering from #archiveteam-bs [06:04] Any IRC client recommendations? [06:16] what's your favorite OS desktop / terminal [06:20] HP_Archiv: I haven't seen the -bs discussion, but: I used to care, but now, honestly, I gave up on looking for the "perfect" IRC client, and just switched to thunderbird because it's convenient [06:21] And yes, thunderbird does have an IRC client, a lot of people don't know about that. [06:22] I strictly use Windows [06:22] *** dhyan_nat has joined #archiveteam-ot [06:22] My only concern for an IRC client would be privacy. As long it's reputable/widely used and development is still there, I really have no other preference [06:23] I'm not saying much because I don't know Windows. I'll say in general, you'll use the first one unless/until you find something you can't stand. [06:24] So something like HexChat or mIRC.... thoughts? [06:24] I heard mIRC is nagware now, idk for sure myself. [06:24] If you already use thunderbird for email, I suggest using its built-in chat option. It has conversation history for each channel like you're looking for. I have never used mIRC, but can recommend HexChat. [06:25] I don't remember whether hexchat can access history from inside the application, but I'm pretty sure it keeps logs [06:25] in the msft store, the free trial is full version. [06:25] No. I use Gmail for email. All of this just to participate in the #archiveteam chat...? [06:26] $10 version is $10 donation [06:26] for hexchat [06:27] Windows: Famously mIRC. It's the widest used, actively developed, 24 years -- It is shareware, only 32-bit, single thread. Second option I'd suggest is AdiIRC, modeled after mIRC, supports mIRC's unique script language, technically better than mIRC. Hexchat is popular, pretty lacking, based on Xchat. I understand that IceChat is still being developed, hobby developer. [06:27] which is simple to setup? [06:28] Freenode channels to join: ##mirc #adiirc #hexchat #icechat [06:28] no irc client is simple to set up :) [06:28] webchat ;) [06:28] heh [06:28] Most IRC clients are simple to "set up", in the sense of install, run, join network/channel, but you'll probably fiddle with the settings for years [06:29] I gave up on doing that and am now using mostly default thunderbird settings. [06:29] Thunderbird still around is it? [06:29] All I wanted to do was to further participate in the #archiveteam / #archiveteam-bs chat to better keep tabs on previous messages/history [06:30] eg: scroll back if I sign out for a day and see missed messages [06:30] HP_Archiv: Maybe you should consider IRCCloud or Mibbit or Matrix or keep using that basic webchat you're currently on [06:30] Ah, if you mean messages that were said while you were logged off, that an inherent limitation of the IRC protocol. You can buy a service like IRCCloud that stays logged on and you can join and look through the messages at any time [06:31] Wait, does IRCCloud have a free plan? I need to look that up [06:31] yeah, if you want persistence, you'll have to generally pay a subscription to someone, or install ZNC on a remote terminal you can access, or install Wechat or Irssi on a terminal someplace [06:31] @Mateon1: yeah that's what I mean. [06:31] I don't think they do [06:32] If persistence were free, then IRC would be filled with 10 billion trillion sock puppets [06:32] Hm, well that just sucks, heh. No, I'm not paying for an IRC chat client... think I might just stick to the web client for now. I appreciate the input though [06:32] https://www.irccloud.com/pricing [06:32] just use hexchat [06:32] or adiirc [06:33] it doesn't matter really, there's lots of free ones, but webchat is the worst [06:33] technically, i think adiirc can be called superior in pretty much every category. i just have trouble having used mIRC for 22 years [06:34] I'll check out mIRC since it's widely used [06:34] Thank you ^^ [06:34] Kinda wish I spent more time on a terminal so I could really enjoy Wechat [06:35] If I'm stuck in a terminal I just launch irssi, but then I always get stuck in the configuration rabbit hole [06:35] heh [06:35] I should really clean up all the dotfiles that are spread out across all my computers [06:36] and sorry, that was WeeChat freenode #weechat [06:38] Whatever happened to Mozilla's ChatZilla client [06:38] did they just roll it into Thunderbird? [06:40] I don't remember ChatZilla [06:40] If it was the XMPP client, then yes, I think so [06:40] Since you both have been helpful with this, maybe you can assist with something else. I'd like to get full access to ingest a site or sites into archivebot with special commands. I was told there is a voice or something else. How do I do this? [06:40] I don't think it was, but I dunno. It used to be FireFox's default suggestion for irc:// links [06:40] in the before times I guess, now. [06:41] You need to be an op to give voice, they are the people with an @ sign before their usernames. I'm not sure what the protocol for that is here. [06:41] What is an op? [06:42] Basically a channel's admin/moderator [06:42] the people with the dead worms next to their names [06:42] Heh [06:42] or maybe a green dot depending on your client [06:43] Okay, well since I have an on-going project with associated websites I want archived and ingested into the WBM (and eventually IA) I'd like to get privileges and become an OP. [06:43] How do I do this and not rely on those with @ in their handle every time I want to submit a site? [06:44] We're probably not the people to ask, then. I'm just doing a small archiving project in between university work. [06:44] Someone like JAA maybe has a good primer for you to dive into [06:45] @JAA: Can you assist? [06:45] Thanks guys [06:45] every site needs a different technique [06:46] Oh, so it's all custom. That makes sense I guess [06:46] the 3 people you were asking for help from should be asleep now [06:46] and to that end, every site usually gets an irc channel of its own. [06:46] so patience goes a long way here [06:46] eg, the death of Yahoo Groups is #yahoosucks [06:46] I don't remember what site you were working on [06:46] I'm very new at all of this, I appreciate you taking the time to provide details ^^ [06:46] but if they said the loaded in AB it does it's thing automatically [06:47] There is also #internetarchive for IA/WBM [06:47] getting voice ops to archivebot sounds like a desirable, but then you have to babysit the jobs you submit [06:48] I was trying to archive the entirety of HP-Games.net which is a Potter-games site dedicated to modding. However, there's a section dedicated entirely to Mods/Custom maps for these games. But each specific map has outlinks to a hosted Google Drive or Yandex download. And not only do I want to archive all of these pages, but also download/capture and re-upload into archivebot. Was told in #archiveteam-bs that this would be tricky [06:49] And I have no idea how to do this ^^ [06:49] HP_Archiv: is the site going under? [06:49] I'll just link as an example: https://hp-games.net/all-mods [06:50] And no, technically not. But it could be slapped with a dmca request at any moment considering the games are licensed under WB/EA [06:50] I understand there are infact client downloaders for Google Drive, Yandex, etc cloud hosting services [06:52] i assume if wget can't slurp them, then you can compile a list of urls to feed to a different software that does [06:52] I guess what I'm picturing in my mind is a near-exact replica of this entire site eventually captured and put on IA, where the download links you see on that page would function just as they do in the live version of the site. eg; you click on a mod map files and the IA prompts for the download of that particular map file. [06:52] maybe. That's an #internetarchive thing [06:52] I know you can also find media downloads in IA for full compilations rather than having to wade through an old half-broken replica of a site [06:53] Hm, okay. The problem is that HP-Games.net doesn't actually host the files. It outlinks to online storage. So I'm trying to port over the site, while reconfiguring how the files are hosted [06:53] with torrent links and the whole shebang [06:53] so it's your site [06:53] Not my site. But I good friends with the person who maintains it [06:53] And yes, do have permission to do this [06:54] I'm good friends* [06:54] i'm not driving at that [06:54] just wondering why you don't swipe his harddrive and upload it to IA [06:54] Heh.. One caveat, he's in Russia and I am in California [06:55] We're part of a much larger HP Modding community on Discord. But he's had this site up for years [06:55] And I'm actually working on a much bigger project than just archiving the site - digital preservation of the early 00's Harry Potter PC games [06:55] Operation Red Sparrow House Elf [06:55] maybe utilizing weather balloons, or owls [06:56] Lol [06:56] *** nepeat has joined #archiveteam-ot [06:56] I would try to keep in mind that the WBM is not a web hosting service [06:56] yes :) [06:57] I'm aware - but #archiveteam creates backups of sites all the time. Why wouldn't this be a valid canidate? [06:57] ideally, only le crem de la Cremlin finds its way there :) [06:58] candidate* [06:58] I mean the project has its own merits - I have several Warner executives and one person out of the Library of Congress helping with the bigger aspect of the project. [06:58] I mean, I dunno. It's just dirty when the site developer is himself involved with finding a way to avoid hosting his own content [06:59] is it? [06:59] isn't it? [06:59] so the scenario is, someone wants to put something in the archive instead of deleting it? [06:59] or am i reading the room wrong [06:59] I disagree. Though I'm not a heavy gamer at all, open-access and copyright issues generally speaking force people to get involved in offshore hosting [07:00] "I don't want to pay for AWS, and Google Drive is being a pain. Host it for me from the wayback machine" [07:00] that was my impression [07:00] Specifically, these particular Potter PC games from 2001-2005 are bound up in a 3-way copyright struggle between JK Rowling, Electronic Arts, and Warner Brothers. And they haven't been remastered. [07:00] IA and Google Drive are not replacements for each other, no, Raccoon [07:00] No, @Racoon, he hasn't uploaded them to Drive. [07:01] Those are different modders/owners of files. Each is a different Google Drive account from various people [07:01] I see [07:01] Or a one-off backup to Yandex as well [07:01] I'm looking to backup the entire site w/outlinked mod files in case of a copyright takedown. Simple as that [07:02] i dont think im qualified to issue an Official Archiveteam Elder Council Opinion or whatever about this today [07:02] * astrid goes back to her hacking [07:02] I don't think official statements are allowed in -ot anyway [07:02] who's the one wearing the snail hat [07:02] dead worm! [07:02] lol [07:02] it's a snail hat :( [07:03] @astrid, since you're an OP, can you assist with this? [07:03] what's up HP_Archiv [07:04] dang, what was the name of that open source cloud storage download manager program I used ages ago [07:04] I want to archive the entire HP-games.net site. But in doing so, example this page, https://hp-games.net/all-mods , want to all archive all outlinked files [07:04] that would be a start for personally yanking all the remotely hosted files to your local machine, to tidy up and organize [07:04] The outlinked files are from various people, hosting files from their own private GDrive or as a backup with Yandex. [07:05] along with a README file and other metadatas, pop in a zip and upload to #internetarchive [07:05] HP_Archiv: this sounds like a big collection that could do with some manual curation from you, so download it all and make it a nice presentable collection of items created with https://archive.org/upload [07:05] what she said [07:06] an "item" can contain one or many files, whatever you think should belong together forever [07:06] That's not a problem. I can do that. However, how do superimpose once the site is archived on IA eg: have the same outlinks point to the hosted collection of curated files? [07:06] Or is that simply beyond the scope of site archiving? [07:06] there isn't a way to link the two nicely. [07:06] technical limitation [07:07] I was afraid of that. Hm [07:08] otrrlibrary.org has, since around 2012, slowly been curating its archive of Old Time Radio programmes into tidy collections, uploading to IA, then simply pointing to that link [07:08] *** dhyan_nat has quit IRC (Read error: Operation timed out) [07:08] Oh well. I'll go ahead and get all of the files together nicely. @astrid can you please ingest hp-games.net and then https://hp-games.net/all-mods URLs so they're archived correctly? [07:08] that's a lot of work and it's nearly midnight and i'm not sober [07:09] This ^^ is why I wanted op voice command [07:09] No worries [07:09] I get you're in CA but assume half the people here are in Europe so you'll get better answers earlier in the day [07:09] markedL: who are you talking to [07:09] He's talking to me [07:10] o [07:10] sorry [07:10] I'm based out of California, actually not that far away from the IA's HQ [07:10] HP_Archiv: here's one such example of what people upload to IA which was previously hosted elsewheres on the web. https://archive.org/details/OTRR_X_Minus_One_Singles [07:10] is OTRR out of copyright in general? [07:10] i personally would have made that item into 150 separate items, but, hey, i'm not them [07:11] and it's sister archive (including more technically superior minutia and archivist materials) https://archive.org/details/OTRR_Certified_X_Minus_One [07:11] @Raccoon, thank you. This is a good example [07:11] the latter example has a much more structured directory [07:12] astrid: seems to work in the IA player, and they did it in 2008 [07:12] So we, the community helping me with this massive project, have gone ahead and put together a 'working archive directory' for these games [07:12] https://archive.org/details/archive_browser_directory_structure [07:13] https://archive.org/details/@harry_potter_games_preservation_project [07:13] It's a work in progress *shrugs* [07:13] so here's another collection that i curated ... i took these 25k pdfs and put one per item, and got a collection created by IA admins to hold it and make it look all nice https://archive.org/details/etsi_standards [07:14] markedL: OTRR is not out of copyright until another couple decades, unfortunately :( [07:14] i need to freshen the collection sometime, there are a few thousand more pdfs published by them now [07:15] Good examples ^^ I'd like to get the polished collection-look you have going there. Any critiques on the pages I linked that you guys can offer up? [07:15] markedL: audio works have now been reformatted (as of 2019) to 95 years +5 for the era of these shows [07:15] (Bearing in mind, I know it's late where most of you are, heh) [07:16] (OTR is also kind of unique, because the licenses are often so spread out thin, that there's really no single rights holder to bully people around, and also weird things to do with federal funding) [07:17] ((things aired for the armed forces, the mutual broadcast system, etc) [07:17] I'd welcome any critiquing for the directory I've put together thus far ^^ [07:18] HP_Archiv: btw, you will want this Windows program. VoidTool's Everything @ https://www.voidtools.com/downloads/ [07:19] you can't curate an archive without it, not really. [07:19] Open source? [07:20] I think yes, or maybe just an API [07:20] I've been using it for a decade at least [07:20] And how does that differ from, say: Invenio, Islandora, Haiku, Samvera, Dspace ? [07:20] I never heard of any of those 5 [07:21] https://en.wikipedia.org/wiki/DSpace [07:21] I come to the table, here, in framing everything in an academic context [07:21] If you have something better than Windows Explorer, you're already ahead of the game [07:21] These are open-sourced software that are used in academic/institutional settings [07:22] I'm just talking about walking your own filesystem and directory structures with a meaningful interface [07:22] Agreed, just surprised nobody knows of these in here. [07:22] robust filtering, regular expression patterns, etc [07:22] bulk renaming [07:23] Hm, okay [07:23] I guess I'm looking to do this in this most proper way possible, and use the same tools that an institution might (regardless of obvious overkill) [07:24] find all folders that DO have both a .PNG and a README file?: child:.png child:README [07:24] stuff like that [07:25] I get what you're saying ^^ [07:25] No criticisms or constructive critiquing on my the directory I've put together? [07:26] And yeah - they are only computer games. So I can hear the eye rolling, heh. But if people *really* want things to be around indefinitely, you *really do* have to dedicate an awful lot of time and resources to doing things correctly [07:26] sorry, I must have missed what you were asking. You mean your upload from August 25th? [07:27] No worries, and yes ^^ [07:27] computer games are culturally significant. especially the older ones that took skill to build [07:27] I'm all over preventing copyright-driven oblivion [07:28] I'm glad we share the same opinion. Agreed, Raccoon. It's these Potter games specifically that I want to see curated and preserved. We're looking for the holy grail so to speak, the prototype source archive for HP 1, Sorcerer's Stone, from 2001. [07:28] hmm, dang. IA doesn't let me peek inside .7z files [07:28] Ah, I took of care that. You'll want to view the html file [07:29] Again, a lot of still needs work. For some reason, the Google Drive export create a double-copy of each files hency while you'll see duplicates in the tree [07:29] oh, i found the nav [07:30] ^^ [07:31] I'd say that looks good to me. Could maybe use more per-directory README files that gives a tweet-length description of every file/.zip [07:32] Or maybe like astrid was saying, smaller clumps, since 14.4G can be unwieldly to the average researcher [07:33] when clumping you don't want to put more than about 10G in a single item [07:33] when you are making it presentable to humans, you want it to be fairly small [07:33] think one of those narrow boxes of loose papers on a bookshelf [07:33] not the whole aisle of books [07:34] I use the walmart plastic tub stratographic filing system, myself. [07:34] There was a meme posted a while back in the Archivist Think Tank FB group... read something like: First person, 'But nobody reads the damn labels..." second person, But I have so much knowledge to give!" [07:35] :) [07:35] These are good suggestions though [07:35] Thank you both ^^ [07:36] Hopefully you can forget all the nonsense I made up on the spot [07:36] *** dhyan_nat has joined #archiveteam-ot [07:36] So I've done other work, unrelated on family films. If you think 10 or 14GBs is a lot... https://archive.org/details/taloccifamilyfilms [07:37] When you digitized moving image film correctly into DPX and TIFF sequencing, file and size add up pretty quickly [07:37] I've delved into sharing terabytes, locally / regionally anyway, it's not necessarily pretty [07:38] file number* [07:39] No, it's not. And I keep hearing that .DPX is too cumbersome a format for film preservation. Because the repo's are absolutely huge [07:39] if I could magic one thing into existence, it would be a form of offline-journaling filesystem for syncing two drives together for magical collaboration [07:39] built into the FS [07:39] That directory of family films I linked above, uncompressed is just under 2TBs, for about 40 minutes of Super 8 film from the 70's... [07:39] HP_Archiv: we try to keep items under 10g because IA can't split a single item across multiple disk drives. so the bigger items get, the trickier the job of the auto-balancing disk allocater gets [07:40] and keep in mind that yes if you make an item bigger than a normal drive it'll cause issues [07:40] 10g is big enough for most uses, and not a hard limit [07:41] Well the thing is, @astrid when you have 40 minutes of film, right, and in film preservation you have each frame representing a single file, a scan.. you can end up with hundreds of thousands of files. In my case, about 90,000 DPX and 90,000 TIFF files [07:41] also, if these are "old games" for "old systems", then probably 4GB is too large [07:41] yeah sure i'm not saying it's bad [07:41] i'm explaining why i said 10g [07:42] Oh I know, I'm just furthering explaining why those films are so large in upload size. [07:42] i'm not stupid, i can see that it's a frame scan [07:42] Digitization is not the catch-all solution for film preservation, clearly... you can re-fresh millions of files every so often when you have one particular film resulting in, say, 30TB's of DPX files... [07:43] can't* refresh* [07:43] For those films, I actually handed over a physical drive with copies of the data to the IA vs uploading over the internet. It wouldn't have been practical to upload. [07:44] But I had the films scanned in 2K, and in theory Super 8 can be scanned even higher in 5K. Could you imagine the size output? Incredible [07:45] are TIFFs necessary anymore? [07:45] Good question. Depends on your needs. I had the scanhouse I used scan to both DPX, TIFF and ProRes files simply because they offered it and, at the time, I thought that the more varied formats the better [07:46] are they uncompressed TIFFs or LZW compressed at least? [07:48] I wonder if your scan house would merely convert the single scan to the other 2 file formats, or if the guy just punched the machine to scan for 2nd and 3rd times [07:48] Everything is uncompressed. And actually, the technician at the IA who helped me with this, didn't upload correctly. The films there are in 2 reels. So for each, I had the scanhouse scan both reels in DPX, both reels in TIFF, and both to 2 ProRes files [07:49] oh gee, so you have 3 scan passes [07:49] He uploaded reel 1, in dpx and reel 2, in TIFF, when in reality, he should've uploaded R 1 & 2 in DPX and the a second copy of each in TIFF [07:49] Yup [07:49] I wonder if that opens up any further opportunity for remastering [07:50] image stacking [07:50] Yeah, it does. And that's a really great question @Racoon [07:50] DPX is, by right, supposed to be edited in Final Cut for example. I simply don't have the computer resources to do that sort of mini-movie remastering, nor that technique or skills [07:50] "two out of three scans don't have that hair right there" [07:51] Heh ^^ [07:52] I was a dipshit and brought a brought several rolls of 20 year old 35mm to be developed... at Walmart [07:52] I spent the better part of the past decade trying to do all of this, self-taught. A lot of time researching, reaching out to institutions and fielding answers from credible/helpful people in film preservation [07:52] The minimum wage employee seemed non plussed about blackening all of them [07:53] It came down to http://cinelicious.tv in downtown LA vs https://www.gammaraydigital.com/ in Boston, Mass. I chose GammaRay because they were a bit cheaper and I lived in Boston at the time as well. [07:55] Both have big-name clients but GammaRay was great. It was a pricey experience. But ever since I had this done 2014, I've been re-migrating data. All those files. So in an effort to open-source old family films, I donated a copy of everything (well almost everything, as explained by that goof-up the IA technician did) to IA and took some of the pressure of re-migrating data every so often [07:56] *shrugs* I was born in the early 90's. These films are of a parent and siblings and my grandparents. When I say I come to the table wanting things to be done correctly, in an academic context, I wasn't kidding :) [07:56] Do you have the negatives? If so, you can always get them done correctly [07:57] yes. the negatives are developed and I put it under a 20k dpi scanner on the university. there's a vague shadow of the former images [07:58] did you ever make a modest useful cut of your films, so they can be shared on social media [08:00] Yes, aside from viewing on IA video player directly, I uploaded the ProRes/Quicktime files (which are smaller in size at around 30GBs each) to a YouTube channel [08:01] So the films can be viewed easily. It's the raw, master data that a pain to keep alive [08:01] indeed. [08:02] I have various backups in place. But in any archives work, the question that anyone in working in this field will bring up is, 'What are saving at the end of the day, and how many resources can we put towards saving it?" [08:02] hell, it's a pain to keep simple cans of film if Universal Studios is any indication. Don't worry about oxygen and rot, they always catch fire long before they turn to dust [08:02] These are old family films from the 70's. But they are a window into my lineage, a window into the past. And I can't put a price tag on that. [08:02] That was a travesty ^^ [08:03] and if they don't catch fire, the BBC will melt them for the war effort [08:03] UMG needs to be held liable, but to what extent I am not sure [08:03] have the BBC melt them down :) [08:03] Lol [08:04] The missing Who episodes ^^ [08:04] guess who keeps rescuing old culturally-significant media? that's right, pirates and bootleggers [08:06] Actually, the Super 8 film that is the source material for these films I've shared is in good shape, thankfully. No brittleness or dedegration before I came along. I sent the films to a place in Hollywood that uses commercial grade film cleaning machines. They're called Lipsner-Smith, the brand. A lot of major Hollywood studios will implement these machine [08:06] If it still exists, it's probably because it was stolen before it could be destoryed. [08:06] That's a great footnote, agreed [08:07] In sort of the same context, getting back to the HP games for a moment, you might've noticed on the main page a mention of the 'HP 2 Prototype' - this only exists thanks to a former developer who disregarded an NDA from EA and kept a local source directory on a personal, non-work machine [08:08] I'm waiting for the first 100% digitally authored book to become completely destroyed due to DRM and the expiration of license keys [08:09] We wouldn't have had the Chamber of Secrets prototype source files if it wasn't for this person. And hence why we're after the first-ever Potter game's prototype files. It's been 17 years, and highly unlikely it exists at all [08:09] nice [08:10] But I and others are in communication with several WB executives, one former head of licensing who oversaw the Potter game development specifically, who expressed interest/piqued curiosity regarding the nature of the project, since it's altruistic [08:10] Yes ^^ [08:10] I shared files from old game betas (c2001) on old drive backups, that helped people recreate parts of the Asheron's Call game, which there is now an emulation server [08:11] If you follow WB you might have heard about the particular outrage over this game [08:11] The person out of the LoC is in the audio/visual department, where they have a modest workflow for video game preservation. Quoting him, 'The Library of Congress is very much interested in participating in conversations regarding acquisition of video/computer game prototype or development-related digital assets.' [08:11] WB went on a mad tirade trying to get the emulator shut down and the developers arrested [08:11] I have not heard this, but it's not surprising. We have been walking a fine line for who to reach out, and who not to. [08:12] but then there are success stories like Subspace aka Continuum that even made its way to Steam. "Illegally" [08:13] If you ever hear of HP Sorcerer's Stone PC game proto/dev file or anything related to development files, please get in touch. On that IA page for the Potter games there are emails there should you ever hear or see anything in passing. [08:13] :_ [08:13] :) * [08:13] That game is split at least 4, 5 or 6 ways and nobody really knows who owns it after a dozen mergers and property splits [08:13] ^^ this [08:14] Virgin Games probably has a nugget of license [08:14] in some tiny foreign country [08:15] So the first 3 HP games were contracted out for dev under EA, but through a sub-company called KnowWonder. Which eventually folded, was acquired by a company called, Amaze Entertainment, which folded too, and was bought out by a now-defunct Foundation 9. [08:15] I'll keep my ear out [08:15] The copyright licensing angles are such a bitch, it's not even funny [08:15] Thank you :) [08:16] One of the funny things about SubSpace is that part of the networking side it got sold to... I dunno what order... FAST, KaZaA, and Skype (eBay/Microsoft?) [08:16] Somebody owns the peer connect tech at least [08:17] doubt anybody remembers [08:18] This is thing - these game studios seem to all be plagued with the same fate of closure/being gobbled up by a bigger/flashier studio or just being dissolved altogether. [08:18] And the company assets, which would include things like game prototypes and development files, have to be hosted somewhere. I can't imagine they just mass-delete things [08:18] And here's the newest thing in Video Games being erased due to license. Watch this video and you might cry yourself to sleep. https://www.youtube.com/watch?v=RTkxzQDo0ng [08:19] oh. they don't mass-delete things. they just stop paying for the hosting these days [08:19] or they close the offices and the contents are denatured and auctioned [08:20] It's not like Sierra Online ever saved the source code to Kings Quest and Leisure Suit Larry. It was core devs that brought the disks home with them. [08:21] Watching that vid now. And the thing is - the latest in our communication is from a former dev who worked on both HP 1 and HP 2. I just received the latest email from him last night, which I can share with you herem [08:21] That's exactly the case with the HP 2 dev - he brought the files home with him, a perchance scenario [08:21] Here's what this dev said yesterday, with name redacted for privacy of others [08:22] "here’s not one single prototype build. That trailer TCRF is comparing against used footage from several different versions of the game from the last few months of production. If you’re looking for a single snapshot of the depot from that time, or near the end, then what I had would have worked for you, but again those discs were destroyed. However you might try contacting name-redacted, who most recently worked on Blackout Club [08:22] Oh, well that didn't copy. Damn IRC... [08:22] You have to consider that PUBLISHING companies don't care about retaining assets for cultural significance. If a product is floundering, like it's grossing below $5 million a year to keep alive, then you send it to the chop shop to be divvied out to anyone willing to buy bits and pieces [08:22] "There’s not one single prototype build. That trailer TCRF is comparing against used footage from several different versions of the game from the last few months of production. If you’re looking for a single snapshot of the depot from that time, or near the end, then what I had would have worked for you, but again those discs were destroyed. However you might try contacting name-redacted, who most recently worked on Blackout Clu [08:22] the same way you sell old car parts on ebay [08:22] "If you’re looking for a copy of the entire repo... I suspect that’s long gone. That would have been on a server at Knowwonder / Amaze Entertainment, but that company no longer exists. You’d need to find someone who worked IT, and someone who has a working copy of the source control software we used back then." [08:23] Yeah... I am realizing this ^^ [08:24] I installed hexchat. relatively easy to get started [08:24] The trailer he's talking about was a promo that was previewed at an E3 in 2001. But there were major changes to the game prior to it's release, but post-trailer of that demo. And we're after the entire repository. [08:25] markedL: there goes another 20 years of your life, poking and tweeking :) [08:25] And again, just on a perchance scenario, the HP 2 dev, different person, had that entire repo for HP2 plus the entire source package for Unreal, all saved on a local disc [08:25] so the best IRC client must be the one with no config options [08:25] The odds that HP 1's proto repository exists are slim, but I still have hope, heh [08:26] Anyway... @Raccoon it's been very nice chatting. Thank you for this ^^ [08:27] HP_Archiv: certainly. hopefully more bits of media and resources continue to reveal themelves [08:27] Gotta keep the faith ^^ [08:27] I do recommend that youtube link though. it was very enlightening, I had never even considered this was a practice. https://www.youtube.com/watch?v=RTkxzQDo0ng [08:27] Oh right, thanks. I had paused it to continue chatting. I'll watch now [08:28] ++ [09:13] *** dhyan_nat has quit IRC (Read error: Operation timed out) [10:03] *** BlueMax has quit IRC (Remote host closed the connection) [10:03] *** BlueMax has joined #archiveteam-ot [10:04] *** BlueMax has quit IRC (Remote host closed the connection) [10:05] *** BlueMax has joined #archiveteam-ot [10:19] *** tuluu_ has quit IRC (Quit: No Ping reply in 180 seconds.) [10:19] *** tuluu has joined #archiveteam-ot [10:46] *** icedice2 has quit IRC (Leaving) [11:36] if someone could manage the YouTube archiving better than me in #youtubearchive that would give me more time to write the code that needs to be written [11:36] I threw in a lot of channels and some of them don't really need to be archived [11:38] channel popularity is not the best metric as it doesn't discover the more interesting stuff, easier to find with searches and digging through related channels [11:38] *** BlueMax has quit IRC (Remote host closed the connection) [11:38] *** BlueMax has joined #archiveteam-ot [11:46] *** dhyan_nat has joined #archiveteam-ot [11:54] *** BlueMax has quit IRC (Remote host closed the connection) [11:54] *** BlueMax has joined #archiveteam-ot [11:55] *** BlueMax has quit IRC (Remote host closed the connection) [11:56] *** BlueMax has joined #archiveteam-ot [12:04] *** BlueMax has quit IRC (Remote host closed the connection) [12:05] *** BlueMax has joined #archiveteam-ot [12:07] @ivan, I have a very long list of YouTube channels, academic/science based. But would the archiving process simply use YouTube-dl or do you have access to the original videos through the API? [12:07] HP_Archiv: I have access to whatever youtube-dl grabs, which is a muxed webm or mp4, and it is usually good enough [12:08] I kinda doubt youtube serves original-quality through their API [12:08] they had this for a subset of videos many years ago as a public format [12:08] @Raccoon I did watch the video on Driver San Francisco. That was just great, thank you for sharing 'creators who aren't capable/worthy of curating their own art' is fitting for this overall issue of game preservation [12:08] HP_Archiv: join #youtubearchive and !a everything soon after the # of tasks drops [12:09] *** BlueMax has quit IRC (Remote host closed the connection) [12:09] *** BlueMax has joined #archiveteam-ot [12:10] @ivan, okay. I sort of conceded a while ago that while YT DL does a good job and all, it's not *actually* the original videos, obviously. And when we're talking about *actual* digital preservation, you go to the source material. What YouTube-dl is doing is not grabbing the source material, by an edited version of it, sadly. [12:10] * Raccoon tries to convince ivan to switch to mp4/m4a then re-download and replace all his .webm junk ;) [12:10] Just want to make sure you're aware. ^^ [12:10] _actual_ digital preservation [12:10] but an edited* [12:10] *** BlueMax has quit IRC (Remote host closed the connection) [12:11] *** BlueMax has joined #archiveteam-ot [12:11] *** dhyan_nat has quit IRC (Read error: Operation timed out) [12:12] HP_Archiv: I had to make a rationalization a long time ago, that I cannot possibly preserve every subatomic particle, and that sometimes 2500 kb/s at 1080p is 'good enough.' [12:12] Raccoon: I agree that .webm is junk but it does allow me to save more stuff [12:12] Beginning of the year, first learning how to use Youtube-dl, I pinged one of the main devs. Asked about this very question 'Is it really grabbing the original videos uploaded by the user?' - No. I then gave an example of film preservation. If you were going to preserve 'Gone with the Wind' in 2K or 8K resolution, you would not pull from a copy of Blu-Ray master. [12:12] and certainly better than nothing. at the very least, I'll have the world's best collection of "thumbnail videos." [12:13] ivan: and indeed, better than nothing. [12:13] You would source the original 35mm or 70mm physical film print that sit in the vault at MGM [12:13] well, you wouldn't do that either [12:13] Lol @Raccoon, yes, yes [12:13] Agreed [12:14] It's better than nothing is what I came to as well, not that long ago for YT DL [12:14] i mean, material sources are great, but you do not possess the talent of the 40 people who worked on directing, printing, cutting and editing the final cut [12:14] you could spend 100 years locked in a room with that spool of film and never come close to "Gone With The Wind" [12:15] If you really want to split hairs, yeah, you're spot on and correct [12:15] while I'm sitting pretty with my 2500 KB/s rendition ;) [12:15] there's always a balance [12:16] ^^ I posted thread on r/Datahoarder about this back in March. And the prevailing viewpoint on there among many - not all - is that YT Dl is 'digital preservation' [12:16] The short answer is yes, but the real answer is no [12:17] *shrugs* If it's as good as we can get, I agree, then that's all we have to work with. I just don't like this aura that surrounds it that YT Dl is the answer to preserving YouTube videos, heh [12:18] it's also hard to tell after you finish shooting Gone With The Wind in a hot sweaty desert, just exactly what material is worth loading on the truck and saving, what wardrobe should go back to the local wardrobe-for-hire folks, what should be sent back to the studio, or what will fetch a price with the Smithsonian museum [12:18] Anyway, sorry @ivan didn't mean to steal your fire. Thanks for sharing the youtubearchive link [12:19] Lol @Raccoon, again - I can't disagree with any of that. All valid points [12:19] It begs the question, 'Why?' for anything we regard as worthy of preservation, indeed [12:20] don't need @ to highlight on IRC [12:20] it's VERY hard to tell in advance just what will strike a chord with the audience, let alone re-write a nation's culture. [12:20] Thanks, didn't know ^^ [12:21] And yeah, that's why the proverbial view of 'we can't save everything' always bugged me [12:22] A fine example is the simple (classic, great) movie 3:10 to Yuma. Given the completely unrelated circomstance of Cuban fascism and the US embargo, it turns out that 3:10 to Yuma can become Cuba's most culturally significant film adored by all, if for no other reason than it's the only good film they had to watch for 30 years [12:22] So culturally significant that Cubans now call Americans, "Yumas" [12:22] Good example of cultural begetting culture ^^ [12:23] culture* [12:23] A small town in Arizona is now the name of the peoples of a nation [12:24] all because a black and white film played on rerun like M*A*S*H [12:25] There is no limit to cultural output, it's limitless. In its rawest form, culture is a singular idea, a thought [12:26] Each thought or idea its own thing, etc. [12:26] But will we ever recover the hats they were wearing on the set of that film ;) [12:26] Maybe having promotional stills of the show of the characters wearing said hats would be 'good enough', heh [12:27] Or highly skilled people, like Adam Savage, re-create a particular prop, etc. Oh that reminds me - you might like this Raccoon [12:27] Yeah, he's quite the craftsman [12:27] https://www.youtube.com/watch?v=29SopXQfc_s [12:28] In this talk, he goes on about this wild journey to recreate an exact replica of statue found in The Maltese Falcon [12:29] The stuff that dreams are made of... https://www.vanityfair.com/hollywood/2016/02/mystery-of-the-maltese-falcon [12:30] Using your logic, the replicas are their own thing - which I would agree - also worthy of being asked the question - is this good enough, is it worthy of being preserved [12:31] That talk he gave, though old, is really good. Worth the whole watch :) Good night ^^ [12:43] Actually, just realized that the actual Ted Talk is here, that other video from Fora.tv is edited for some reason, https://youtu.be/rEg-ZNB3qyI night [12:46] that one was pretty good. very on point with the conversation at hand too :) [12:47] *** mls_ has quit IRC (Remote host closed the connection) [12:48] his 20 gig download folder is adorable though :) [12:53] Yeah, I like how world's collide sometimes like that, very on point indeed :) And 'over 20GBs of info' that was from 2008, go figure heh [13:33] *** synm0nger has quit IRC (Quit: Wait, what?) [13:34] *** SynMonger has joined #archiveteam-ot [13:39] *** BlueMax has quit IRC (Read error: Connection reset by peer) [13:55] *** yano_ is now known as yano [14:08] *** synm0nger has joined #archiveteam-ot [14:08] *** SynMonger has quit IRC (Read error: Operation timed out) [14:08] *** HP_Archiv has quit IRC (Quit: Page closed) [14:51] *** phillipsj has quit IRC (Remote host closed the connection) [14:51] *** phillipsj has joined #archiveteam-ot [15:39] *** jmtd has joined #archiveteam-ot [15:41] *** akierig has joined #archiveteam-ot [15:41] *** akierig_ has joined #archiveteam-ot [15:46] *** akierig has quit IRC (Read error: Operation timed out) [16:11] *** akierig has joined #archiveteam-ot [16:17] *** akierig_ has quit IRC (Read error: Operation timed out) [17:18] *** manjaro-u has joined #archiveteam-ot [18:09] *** DogsRNice has joined #archiveteam-ot [18:35] *** X-Scale has quit IRC (Ping timeout: 252 seconds) [18:35] *** X-Scale` has joined #archiveteam-ot [18:36] *** X-Scale` is now known as X-Scale [18:41] *** akierig has quit IRC (Remote host closed the connection) [20:04] Does anyone have a sort of exhaustive list of various time and date printed format variants, and how they might classically have been used, or whether they have an ISO standard or conventional name attached to them [20:11] *** akierig has joined #archiveteam-ot [20:16] *** ShellyRol has quit IRC (Read error: Connection reset by peer) [20:17] *** ShellyRol has joined #archiveteam-ot [20:18] *** X-Scale` has joined #archiveteam-ot [20:19] *** X-Scale has quit IRC (Ping timeout: 252 seconds) [20:19] *** X-Scale` is now known as X-Scale [20:31] *** X-Scale has quit IRC (Ping timeout: 252 seconds) [20:44] *** mls_ has joined #archiveteam-ot [20:49] *** X-Scale has joined #archiveteam-ot [20:51] *** dhyan_nat has joined #archiveteam-ot [20:56] *** akierig_ has joined #archiveteam-ot [20:57] *** akierig has quit IRC (Read error: Operation timed out) [21:47] *** dhyan_nat has quit IRC (Read error: Operation timed out) [22:28] *** BlueMax has joined #archiveteam-ot [22:42] *** DogsRNice has quit IRC (Ping timeout: 252 seconds) [22:42] *** akierig_ has quit IRC (Quit: later_gator) [23:23] *** wp494 has joined #archiveteam-ot [23:42] *** foureyes has quit IRC (Quit: brb) [23:44] *** foureyes has joined #archiveteam-ot