 Thank you so much for having me and especially thanks to Diane and Cliff for making it work to present remotely despite the in-person conference. I really appreciate this opportunity to share what we've been doing with Succo this last number of months. So I'm Quinn Niembrowski. I'm an academic technology specialist at Stanford in between the libraries and the division of literatures, cultures, and languages. My background is in Slavic linguistics, medieval East Slavic linguistics, which is a laugh line in like half my talks, but very strangely, it's ended up, you know, this year, been the year where kind of all of these pieces that I've worked on, you know, from digital humanities to the medieval Slavic, you know, to technology of various sorts and working in central IT for a number of years all came together through this project. So saving Ukrainian cultural heritage online is a group of over 1,300 friends and new friends mostly from North America and Western Europe who came together in the wake of Russia's invasion of Ukraine in February to try to protect Ukrainian digital cultural heritage. To be honest, we didn't really know what we were getting into when we started this, but as we worked on the project, you know, we discovered how quickly, you know, ever since World War II there are many organizations that exist specifically for the purposes of protecting cultural heritage. You know, there's Europe or Nostra, there's the Smithsonian that's been involved for decades, but there are very, very few players in the space of protecting digital cultural heritage. And over the last 30 years, the amount of effort that has gone into creating digital cultural heritage through digitization of physical objects, to websites, to born digital things is immense and these are materials that need, you know, care taking during a war as well. So here's how it all started. It feels sort of sad and ironic now, you know, how much of SUCHO, especially in the early days, depended on Twitter, the academic scholarly communication infrastructure that, you know, is now on fire metaphorically, at least if not literally as well. Anna Kias, who is a music librarian at Tufts, posted on February 26th, just a couple days after the war started, she wanted to organize a data rescue session focused on music collections. She was at a music conference and that was kind of her immediate community of folks that she could bring together to, you know, archive websites that had a fair amount of music content. And Sebastian Maistorovic immediately replied, familiar side of the world at the Center for Digital Humanities and Cultural Heritage in Vienna, Austria, suggesting that WebRecorder might be a tool that might be useful for this context. And I saw this and we all got together and had a meeting the following day. And just a couple days later, we launched SUCHO, where Sebastian had paid out of pocket for server space to be able to start archiving sites around the clock using the WebRecorder suite of tools. And we, you know, as one does, we put out a Google form for people to sign up to help volunteer. And all of a sudden, you know, we were being flooded with volunteers, you know, over 400 people signed up to help in the first, you know, a couple days, even as we were still setting up the Slack. We were writing tutorials. Honestly, I had never really done web archiving before this project other than sending things to the wayback machine. And so, you know, we were learning these tools ourselves for the first time, writing documentation for the next people who would be learning them, you know, minutes after we ourselves had figured out these things. This was our web archiving and collaboration toolkit. We were gathering links from all kinds of different sources. You know, we were using Wiki data. We were having people send us links. We had a link submission form. And we would send things in two places. Everything would go to the wayback machine. We would check to make sure that it had been archived before and we had volunteers go through and make sure that the capture was thorough and caught everything, even the sub pages. And for things that were missing, we would have them submit new seed URLs to the wayback machine to make sure that those are captured as well. We were also having people use the web recorder suite, which is available as a browser-based plugin as a Docker container. And eventually, we were the first prototypers of Browser Tricks Cloud where people could do automated web archiving in their own browser. For a lot of volunteers, we had a number of volunteers from, you know, libraries, archives, museums, universities who had some experience with kind of cultural heritage technology. But we also had a lot of people who had never tackled that kind of thing before or similar to me had only used the wayback machine. And so our tutorials and video guides and hands-on sessions were used to train people to install Docker and run things on the command lines sometimes for the first time. And it was amazing how with a little bit of coaching and support, people at all kinds of kind of self-assessed levels of technology were able to pick this up and work with it. We were communicating around the clock via Slack. We had a free Slack instance and we got an extension of our free trial actually until November. And that has been our hub for communication, you know, synchronous, asynchronous around the clock. Sometimes we would hold Zoom sessions, but honestly, most of the project communication happened on Slack. And Google Sheets was, you know, perhaps the unsung hero of this project. We had this gigantic spreadsheet with multiple different columns for all the links that we were getting in. We would set the status there. We would have information about where it was hosted. Initially, we were prioritizing sites that were physically located in Ukraine with our primary threat model thinking of, you know, missiles and physical destruction of servers. It turned out over time that we stopped doing that and we would take any site that was related to cultural heritage, even tangentially. You know, we had everything, not even just libraries, archives and museums, but also, you know, children's after school programs where they were learning Ukrainian language and cultural heritage, you know, train museums for, you know, teenagers to learn to be train conductors dating back to the Soviet era, even fanfic sites. Anything that are, you know, that represented places where people were engaging with Ukrainian cultural heritage. We wanted to make sure there was at least one or two safe copies of, you know, distributed, you know, using our browser tricks tools and also in the Internet Archive. It turns out, as I was saying, that, you know, physical destruction was actually not our biggest problem. It turns out that of the sites that are hosted outside of Ukraine, refugees are not great about paying their server bills. Completely understandable. There's a lot of priorities. So there were sites that were hosted, you know, that were not in any sort of physical threat that went down due to non-payment. And so we wanted to make sure that we could capture as many of those as possible as well. We did some creative things with trying to find sites. You know, in addition to talking with subject area experts, like the Bavarian State Library, for instance, gave us a giant dump of Ukrainian URLs that they had curated. We also went to WikiData, but it turned out that a lot of things that were in WikiData were out of date or the domain names had lapsed and the sites were now, you know, gambling rings and things like that. Sometimes more audacious even. So we had to go look for a lot of sites. We didn't know what all was out there. And we had volunteers, especially those involved in situation monitoring, like Erica Peasley, who would look at the arrayed alerts that were going out on Telegram and then look within those regions, actually walk through the streets using Google Maps of cities under attack, looking for the cultural heritage icon and then finding, you know, do those places have a website? And if so, she would add it to our list. The situation monitoring also fed into our prioritization, where generally we just sort of went down the list with anything that hadn't been flagged as a problem or spam or things like that. But when there were active attacks going on, she would sometimes reshuffle the priority list saying, okay, focus on things in this region or focus on things in this city, knowing that those websites might be the only, you know, remaining traces of some of these institutions that were being destroyed, even as we were archiving their websites. It really was a digital Dunkirk effort. You know, Dina Strong, one of our volunteers who spent, you know, these 12 hour days, you know, taking vacation off of work to help wrangle the community came up with that phrase. And it really is apt. We would crawl on any device that could possibly crawl. And this included even Raspberry Pis. We had a volunteer figure out how to run the browser tricks crawler on a Raspberry Pi device and wrote out some instructions and they're now on our websites. People were dusting off laptops that they hadn't used in years and plugging them in, you know, wiping their hard drives so that there was space and archiving, you know, on those computers. We, for a number of months, stored all of our data on Amazon S3 using some free credits that Amazon Central Europe provided for us. We also had a number of mirrors at research institutions within the US and Europe, including ones that we never really spoke about publicly to ensure that there were multiple copies of this data, even if some of our data sources were compromised. In the end, our goal is to give this data back to the institutions if their websites are destroyed, if the institutions are destroyed, you know, extracting the data and helping them, you know, reimagine and rebuild after the war. So we had to make sure that the data was going to be secure. Our volunteer pool really ranged the whole gamut from children to retirees. One of our volunteers sat down with his five-year-old and to archive a website and he asked his child to draw what they were doing as they were archiving websites. You can see in this picture here. This is on March 17th. And, you know, the success of that experiment that he did, you know, got me thinking, like, I too have young children. I wonder if they could get involved. And of course, you know, I live in Berkeley. It's a very sort of, you know, socially active community. You know, the kids have been talking about the war in school. And so we hosted what I think may have been the first ever web archiving event at a, you know, for the PTA of an elementary school. We had the event over Zoom and, you know, everyone was there with their laptops. I had curated a list of children's libraries for the kids to go through and families got together and were able to help in a tangible way by archiving websites. And this is my now nine-year-old Sam here in this picture. One thing that we discovered over the course of this project is that, you know, key glam infrastructure is really poorly adapted for web archiving. You know, there were, for instance, I now have this fear of calendars on websites because event calendars, it turns out, are essentially crawler traps where they automatically generate links to past events and future events even when we're talking in like the best distant future or the best distant past where there are absolutely no events. The calendar will continue creating pages and web crawlers will keep going through those pages and not actually get to the content of the site because they're too busy time traveling. Itavis is a major library catalog system that has been widely adopted in post-Soviet countries developed in Russia. And it is a nightmare to crawl. I mean, you functionally can't. And so what we ended up doing was one of the grad students at Stanford, Yogi Karakov, wrote a web scraper to be able to deal with Itavis library catalogs. And these were really essential because when libraries are damaged or destroyed, we want to be able to have records of what their holdings were before the war, both to try to get that data back to them. If we can find copies of these books or have them donated from people in the US, professors with libraries, or figure out a digitization strategy to get them some copy of the book in some form back. But this had to have a special scraping method to be able to get that data. As did dSpace, unfortunately. While there was not a lot of adoption of some of the other widely used standards like tripleIF, for instance, in Ukraine, we were actually very happy to discover that there were not tripleIF servers that we could find where we needed to back up that data. But dSpace had been fairly widely used for academic publishing, for archives and libraries and galleries. And it is incredibly hard to web archive dSpace with the tools that we have. So, some food for thought for future development, I think. We have the Sutro Gallery and Equipment Fund. At some point, we realize that archiving websites is all well and good, but it's limited in its impact when there's so little of the country's cultural heritage that's been digitized. According to the Ministry of Culture, less than 1% of Ukrainian cultural heritage holdings have been digitized. And so we've been raising funding and working on getting equipment to Ukraine. We have partners there who are developing tutorials in Ukrainian for how to use these materials, how to upload things. We've developed the Sutro Gallery, which is an OMEKA-based platform. If people need somewhere to put these materials so that they'll be visible to help raise funds for these institutions. One of our biggest partnerships right now is with the Cherkasy Regional Library, where they had an art contest for internally displaced children. They wanted to digitize all of this art and turn it into an exhibit that could travel around the world to help raise money to support the community. So we managed to get them a book-eye scanner. It's on its way to them. They have a lower-end Caesar scanner right now. And we're working with them to try to help lay some of the pedagogical infrastructure to be able to duplicate this for other regional libraries and other places that are maybe not in the same high-priority and high-profile level as to receive the kinds of international support they would get otherwise. This is just kind of fun. We also have a meme wall. We realized at one point in the spring that memes are in some ways kind of like, letters home from the war or letters within a community from a war for this particular crisis. And the historians of the future, as much as we sort of see them scrolling in our feeds and we laugh when we move on, historians of the future are going to want this data. And so we've started curating a selection of these memes. Anytime there's a big event in the war, there's always this outpouring of memes and we've been trying to capture those across multiple different languages, add metadata and make it possible for people to browse those. Unlike the websites where we're really seeing this as something that we're doing for the Ukrainian cultural heritage community and that data will be their choice about whether we keep them or get rid of them after they have their data back. The memes we really see as a resource for future historians and already we have students working with this material both in Ukraine and in the US. We are also a part of these conversations about the building of a future National Digital Library of Ukraine, which primarily is a partnership between UNESCO and the Ministry of Culture and Information Policy of Ukraine and the Ukrainian Library Association. But again, Succio has found itself in this weird place of being the people in these conversations thinking about the digital stuff. So it's a multi-governmental effort and so naturally it's going a little bit slowly, but we're looking forward to helping advise on this one in the future, thinking through what the National Digital Library might look like. Already we're talking with Ukrainian Library Association about setting up their own d-space as it happens for their own internal materials as a way for them to kind of get their hands dirty with kind of building archives and thinking about what that looks like to run them in the long term. The situation is actually worse than ever for Ukrainian internet. This is the uptime of the sites that are on our list starting in April and running through now. And there was a while where we were thinking like, maybe we didn't need to do this, maybe actually there won't be any data lost because while there's some degree of fluctuation with any given site on any given day, the sites were mostly online for much of the war. And really it is only within the last few months that this has changed dramatically. And in fact, the risk to the sites that are at least hosted in Ukraine has also increased dramatically because of power shutoffs. So unplanned power shutoffs can lead to data corruption. The risk of data loss is really higher than ever. And I think it really underscores the need for us to not give up and not call us a done effort. This is something that we kind of need to be in for the long run until the community is ready to rebuild. If you're interested in learning more about SUCHO, our website is sucho.org. We definitely need help with a variety of things, including the gallery, including some metadata. We could always use some technical help as we're thinking towards some of these bigger projects with the National Digital Library of Ukraine. And at least speaking for myself, I'm incredibly grateful to the support that I've gotten from Stanford Libraries to work on SUCHO. This was literally all I did for two or three months. And now I'm balancing it with everything else that I'm doing. But the experience of working on SUCHO has led me to get involved in web archiving more broadly within Stanford Libraries and thinking through what the future of that looks like for a variety of different use cases, including web archives' data and sort of preserving these manifestations of cultural heritage for the future. So yes, thank you so much. Thank you for having me remotely. And I look forward to your questions. Now I'll turn it over to part two. I guess I'll stop sharing the screen so that you can just do that. Thank you, Kuhn. Yeah, if, thank you. So if there are any quick questions about this talk so far, perhaps this is the time to ask. She'll hopefully be around till the end of the talk. So if you have more detailed questions, we can address it at the end. I'll be here. We can do the questions at the end if that's easier. Yeah, say it again. You say something? Oh, we can do questions at the end if that's easier. I mean, yeah, if there were any clarification or any small questions, we can perhaps go with that now. Otherwise, we can wait till the end. All right, so paging through this far. Hello. I'm Saud Alam, a Web and Data Scientist of Wayback Machine at the Internet Archive. And at the Internet Archive, we were fascinated by the work Quinn and Team did to kind of establish this hundreds and eventually thousands of people, kind of rallying behind and kind of collecting all sorts of seeds and reviewing stuff and finding all sorts of ways of archiving the web. And Mark Ram, the Director of the Wayback Machine, said, hey, let's go and help them. And find out how we can learn more and how we can help. So basically, many developers from our team me, Mark Ram and some other members of the Internet Archive not necessarily working for the Wayback Machine, they all joined Sucho's Slack and can we help? Sucho actually stress tested SafePageNow service of Wayback Machine by sending a lot of data to us. And the person maintaining SafePageNow service is coming back and kind of, you know, okay, so this is down. Okay, I'm going to fix this. Go ahead. Maybe we can take more load and whatnot. So at the Internet Archive, we are a library and we do, the tasks other libraries do like collecting stuff, preserving, lending it, material, collaborate with other people, provide access to the data we have, make data discoverable by making it indexed in, you know, searchable and whatnot and provide helpful services on top of it. So this is Sucho collection at the archive.org and this is not the web collection. There are all sorts of books and videos and audios, images, you can find many things in here. But then Wayback Machine team started kind of ingesting data, the spreadsheet that you saw before into SafePageNow. I mean, we did not consume it like one entry at a time but SafePageNow comes with many different APIs. You can submit data to it using emails or Google spreadsheet. You can submit one by one. There is an HTTP API that you can use to interact with. There is a browser extension. So many different ways to go about it. The Sucho community, some folks from there actually created an add-on that made the Google sheet-based workflow even better and easier. Then some other tools that members of Sucho team were kind of using to archive the web locally, not using internet archive to archive those. They were creating Waxy files and there are about 50 terabytes of those. They were sitting in Amazon S3. We are in the process of moving that making a copy into internet archives data box. And these are Waxy files, not the traditional work files. So these are actually wrapper around or a container of work files. Our system is not quite ready yet to ingest it as is but we have experimented with that and hopefully soon we'll be able to serve Waxy files through a Wax machine or in other ways. So as we browse the web, we sometimes see things that we realize like, oh, I wish I should have archived that. Please do, right? If you see something, save something. That's the motto of SafePageNow which is a service of web-backed machines. SafePageNow. If you go to web.archive.org, somewhere in the middle there will be a place, there will be a form that will say, save page or SafePageNow, use that. And not just web-backed machine SafePageNow, there are other web archives that are providing services so that you can archive pages there. Please archive in multiple places to have more diversity. Internet archive runs a subscription service called ArchiveIt where different organizations, those who don't want to have their own web archiving program in-house but they do want to have web collections, they can subscribe to our paid service basically of ArchiveIt. And we internally use it sometimes. Even the web-backed machine team uses it. MarkGram created two collections here which is like a news colon Ukraine, news colon Russia. And it's collecting a bunch of seats there and periodically archiving those. And over time we have collected more than 3 million, you know, captures from Ukrainian sites and 24 million plus captures from Russian sites. But perhaps it's not surprising to the audience here that the archiving just a seed list is not enough and archiving the first few hops grows exponentially. Here one seed from cnn.com resulted into 31,772 captures. So how it happens, you get one page and you archive the HTML of the page but you need JavaScript, you need CSS, you need hundreds of images in there and over time there are more and more resources per page. So you archive all those and you get that 338 embeds basically. Then you process that page and you find how many hyperlinks are there. And those external links in this case come out to be 174. Then you go download all those and all those pages will have page requisites like images and CSS and JavaScript and whatnot. And then you get this big number. That's how it expands really quickly. Now archiving regular static pages is easy. We have been doing it for a long time. But how about archiving Facebook or Twitter? Or Telegram or a number of other services. Sometimes we use traditional tools and it works. Sometimes it doesn't work. But when there is a crisis and there is important material there, we try hard to get it. And even if it needs more engineering and specialized tooling around it, right? So it turns out there is a page on archive team which is a volunteer driven archiving community that's been around for quite some time. They basically were up to this task and they went to this platform where there are channels of Russian and Ukrainian and Belarusian and other places. And they got data from there. And they started collecting those specialized engineer ways to download Telegram data from there. And we started ingesting that content with that machine. And this is a dashboard that we have which we see on a daily basis what's going on in our different collections. And so far it's been 22 billion Telegram URLs archived. Then vk.com, which is basically kind of a Facebook of certain country that has its own challenges. And we started archiving that as well. So there are 3K regional groups with daily snapshots since May 2022, about 5 million posts and about 200 pro-war groups with daily snapshots. So we were archiving it aggressively there. Then saving Russian independent media. There are a bunch of independent media sites there and we wanted to archive and preserve that. So we collaborated with Bard College and pen.org on that task and we are archiving those sites as well. And here is a dashboard of that which is so far it's 140 million plus URLs archived in that one. This was about 70, 75 or so initial seed that we got of independent media sites. We don't know which sites are independent media there but there were experts who helped us on the other side. And that's what collaboration at every stage kind of helps. Then we started collecting the data. Now they want to have access to the data that we archived and we have been experimenting with something called collection search in Wayback Machine. So the amount of data that we have at Wayback Machine is enormous and indexing it is beyond our computational capability at this point. We cannot make the whole thing full text searchable. So what we are doing now is like we are identifying smaller collections or things that we think like there is an immediate usefulness to make it available in more ways and make it more discoverable. And we index it full text search, index and make it available. So now if you go to web.archive.org there is a collection search option. You can search for GovDocs in there like end of term crawls that we have made it available. PDFs from GovDocs, those are indexed. In fact PDFs from all the other sites are also indexed there. Poetry.com is indexed, that site is not there anymore. And a bunch of other collections, I think there are like 15 plus collections that are available now and this number is growing. So this Russian independent media collection that we have built that is also now available. Hopefully going forward all the SUCHO data that we have collected that will be indexed and that will be available as well. It's not there yet. So this index that we created, not everyone is going to access it from way back machine interface. So we have created an API for that. Now if you are a researcher and you want to put like a UI around it with some of these collections that we are exposing, you can certainly access that. And this work was again a fruition of collaboration with Media Cloud folks. Media Cloud is an organization that collects news, seed data from all around the world in many different languages. And they give us like a daily feed and we archive those URLs and we index it and make it available. So Archive It has also been supporting multiple partners and some of them are creating collections on Ukraine or Russia or these water related collections. Now we got these URLs and when Wikipedia has become like one of the knowledge sources and URLs die quickly in there, especially in countries that are in crisis. If you go to Wikipedia in Russian or Ukrainian languages the URLs there are more vulnerable to disappearing than some of the other Wikipedia, right? So we increased our activity on those Wikis to kind of fix those links before they break or once they're broken we just change it to way back machine links. And for that we have a bot, we call it Internet Archive Bot that periodically goes and kind of checks all these Wikisites which runs on several hundred Wikisites on a daily basis and fixes links. So what it does basically, right? Here you see there is a references at the bottom of the Wikipedia page. Some of those links if you click on those they will lead you to a way back machine page because the original link was broken it's not there anymore, right? It's 404 or the site is down or something. And here's a little dashboard that we have. So we have recently crossed like 15 million links fixed on Wikipedia, all the Wikipages combined and by the end of the year we'll cross 16 million mark which is another arbitrary number but it's a huge one. So basically we have made sure at least those many million URLs in Wikipedia they will live. The references will stay in readable even if the origin site is down or gone or the content is changed. Here is a monthly activity of our Internet Archive bought on different Wikis and here we're comparing Russian Wiki and Ukrainian Wiki and English Wiki. So in the early days the blue line shows like I mean the English Wiki had more activity than others but towards the end of February 2022 and so so much activity was kind of you know gone into Russian and Ukrainian Wiki and now it's gone, caption to calm down and gone to the normal speed now. So that was the website of things, right? And how about television? We have started archiving television as well in a lot of, we have specialized interface in the Internet Archive when you go to TV news archive and if the news has corresponding captions they will be available. You can browse through it and read through it or interact with it. So this is TV news archive and you see a Fox News thing here with the captions. But Russian TV generally has no captions available so it makes it more difficult to understand what's going on in there just by seeing titles. That's a language that not many people can understand. So this is a TV visual explorer and this is a work from GDEL project where it processes TV news and transcribes the text, the first basically converts the speech to text, transcribes it basically and if necessary uses translation tools to make it more accessible to users not knowing the language and also it creates periodic snapshots like takes thumbnails and puts them into like a strip like thing where you can see what's going on like I have like a visual overview of news even if you don't understand the language and if you click on one of those thumbnails it will land you to the Internet Archive on that particular recording at a specific time so you can watch from there and the transcripts will be available there. It's kind of an interactive system and we are doing it for all the Russian, Belarusian and Ukrainian TV. Now there are with this system in place GDEL has been experimenting with a lot of other projects around this when you have like the TV data collected by the Internet Archive and they can ingest it into their system and do all sorts of machine learning things there. So what kind of value edit services you can provide with that. GDEL's projects blog is an excellent place to follow what's going on in this space. Here is TV Rain, that is a Russian independent TV or shut down for a while and we got disks from them shipped to us. We loaded it, this is Mark Grom's office basically copying data from those disks into our archives and it made it into Internet Archive's data box so now you can browse through and access it to those TV recordings. Now we got that data and we passed it to the same visual explorer we talked about before so now it is more interactive and accessible basically. Then we also archived Echo of Moscow which is like a Russian independent radio. We archived their content, it's not a web archive, it's a regular audio archiving basically or radio in this case. Then we collected Ukrainian books basically, there was a call for that on our blog and people were donating books and they're organizing those collections. Here is something called Better World Books which is like where you can buy used books basically and a cheaper price. That is a subsidiary of the Internet Archive now, a nonprofit, so what we do is every time there is a reference in Wikipedia for a book that we don't have in the Internet Archive, we add it to a special list, which is our wish list basically and every once in a while we send that list to Better World Books and when they find those books on their conveyor belt they basically put it on the side and then we send it over for digitization so that we have those books. So it is our kind of long-term goal of making Wikipedia more resilient and references that are there in Wikipedia that are not web references but books and other non-digital materials. How can we make sure that those references live and be accessible in the future? Then as we showed before, when we have a copy in Webeck Machine of a webpage and if that link is broken, we go and fix Wikipedia page to point to Webeck Machine link or any other archive link basically in non-Internet Archive archives. But since we also have books and other materials we can fix those references as well. There is no broken reference to begin with, just that if there is a book citation in a Wikipedia article, how to go to that book basically? Now sure you can have a library nearby and if they have that book you can go and find that reference but Internet Archive is a library that is accessible so why not link there, right? So we try to identify such references and link to the page number so now if there is a book reference in a Wikipedia article you can interact with the link there and it will open it to a book reader so that Internet Archive has at that page number so you can access those citations that way. So yes, it's a whole involved process. We acquire these books, we digitize those things and we make it accessible in many different forms and we also kind of know where things are cited if we can create a link there, we do that as well. And this is kind of an open problem because some books will have hard references like DOIs or ISBNs so we can have like a one-on-one matching but sometimes the citations will not have that rich metadata available and in that case we need to find more reliable ways to link more books basically. So with that we can conclude and if we can take any questions but I can go through all the things that we did. Basically we supported sucho.org when there were only like a few hundred volunteers in beginning that group or basically organization. We started archiving Russian and Ukrainian news sites and other websites from those countries, archiving telegram, archivingvk.com, saving Russian independent media in collaboration with Penn.org and Bard College. We have many archiving projects that are around war-related sites. Then we are also working on Russian, Ukrainian and Belarusian TV material with the help of GDELT project. Those are excellent folks up there. Then we are also archiving and saving independent TV like TV Rain, independent radio, which is TV Echo of Moscow and we are always looking for more resources that we can archive and make it available in the future. And we are enriching Russian and Ukrainian Wikipedia sites. I think Quinn perhaps can summarize the work on her side and then we can take any questions from there. Yeah, I'm happy to take questions. All right, so any questions? How much time do we have? About 10 minutes? Excellent, so I'm not sure how you're gonna relate this to Quinn, I'm not sure if Quinn can hear the AV here. First of all, I just wanna thank the Internet Archive and everybody who was involved in the SUCHO is a project that was personally pretty important to me. I'm curious about what the lessons learned are. So I remember a few years ago with a CNI event, they were talking about the dashboard that evolved at Johns Hopkins for COVID data. And there was discussion at that event like, well, how do we prepare to present this sort of data or to react to a crisis like this again in the future? Obviously the crisis of this magnitude isn't impossible. We're gonna continue to have data online in this fashion. So what lessons do we learn from the SUCHO or Internet Archive experience in Ukraine that we can carry forward? Sure, Quinn, were you able to hear that? Or do you want me to repeat it? No, no, I was able to hear that okay. Yeah, I mean, it's interesting. I mean, one of the things that we learned very quickly with SUCHO is that for a situation like this, you really need a balance of individuals acting on their own volition and institutions that are ready to step in on a slower ramp up pace. I mean, realistically, you can't expect institutions that are thinking about things for the long term to be able to react as quickly as necessary in a crisis to start gathering this data. I mean, this really is the purview of volunteers who need to start going and doing this. But having organizations like libraries and archives maybe have a plan in place for what to do in this kind of crisis. How do you get in touch with people who are volunteering and doing this kind of work to support them, be it through mirroring or through longer term archival storage or for help of curation efforts? There's a lot of work that needs to be done, both in the short term and in the longer term for institutions and I'm not sure, I'm not sure institutions are necessarily, have been thinking about this kind of work as part of a global emergency preparedness effort, but it might be putting some thought into how you want to liaise with volunteer groups that spring up to respond to these crises in order to be able to support that work and ensure the preservation of the data in a medium and longer term. Yeah, so to add to this, basically, you actually hit the title nicely. I mean, that's what the objective of this talk was. Perhaps we were trying to encapsulate that. Maybe, I don't know how much justice we were able to do, but my personal observation is we really need more Quinn. Honestly. And thank you. And it's not surprising because soon after that, there was crisis in Sri Lanka, just like a small country. And I personally tried to reach out to a bunch of people that I knew that were basically originally from there, they are in the comfort of living here in the US and thought maybe they know what is important for them, what needs to be archived, and if they can help, they didn't step up, basically. And I understand their priorities might be different, but I mean, that's where I think there needs to be more motivated people who know what's going on. Again, and within this year, there are more events. Iran is going through some crisis and what is happening there. We can see some things, we can archive some stuff, but when there are curtains around it, unless someone is inside and somehow the data can be transferred in a reliable way, how are we going to preserve those things? It's an open challenge, it's a very difficult one, honestly, and chasing through. Different communities use different platforms, for example. If you are in India, in some other countries, they're using a lot of stuff on WhatsApp, not a platform very common here, but in some places they will be using Telegram more often than not, or there are some decentralized chat softwares that they use where they communicate and express how to penetrate in a safe way without hurting anyone's identity. Those people are at risk, basically, and we need to be very mindful about it, right? And then archive that material in a meaningful way. We certainly learned a lot of lessons from this, like the kind of infrastructure of preparedness and making sure that things are indexed and made available to researchers as soon as possible. Those would be some things that we can carry forward, but yeah, we need more volunteers, more motivated people. Again, from internet archive side, we are there to help. Yes. Thank you, thank you both for those presentations, and hello Quinn, good to see you. The question I'd like to ask you is, how do the lessons that you've learned apply to how we may respond to climate change? And by we, I wonder, academia as well as nonprofits, like the internet archive. Yeah, that's a hard one. I think the climate change questions really challenge some of the underpinnings of our assumptions about what we have and what we will have and what we will have access to and how we engage with those things. On one hand, absolutely no regrets of what we've done with SUCHO, I think it's essential to have done this work, but in the longer term, I don't think, digitize everything and everyone will have access to it always, everywhere is going to be realistic. I mean, I think thinking about portability of data in ways that don't assume a functioning global internet or the sort of pre-ecos that we're seeing in Ukraine right now. What happens when people don't have power reliably? I mean, I'll admit that kind of on the spectrum of digital humanities folks, like I've tended to be less into the minimal computing. I don't think that we can static site our way out of our impact on climate change, but I mean, at the same time, like these questions are real about what is the future going to look like when we don't have reliable power necessarily or reliable access to data. I mean, I don't have any answers to this other than there's the things that we need to do now, but like we also need to be thinking more creatively about kind of what the future of cultural heritage looks like. And honestly, maybe it's going to be more analog or it's going to be differently digital, portable in ways that we can have access to for certain periods of the day where we have access to power or electronics. I mean, I think there's a lot of interesting and creative things happening in the maker space area with this these days. You're thinking about data in different ways, data physicalization and things like that. The Digital Humanities Center at Princeton has done some interesting things there. But yeah, I mean, I think maybe the big takeaway from SUCHO when it comes to climate change is kind of how far you can get by getting a group of people who pair about something together and to combine their creativity and resources to go do a thing and the things that we will need to do to ensure access to some kind of data, assuming a future of limited power or electricity connectivity and travel are different, but that same kind of human ingenuity and collaboration I think is the same. And at the Internet Archive, I mean, this is a growing concern everywhere. And I think we collectively are still kind of you know, understanding what's going on and what can be done. I mean, a lot of people were in denial more that this thing exists. That is changing, which is a good news I think. But it came with a hefty cost of a lot of people not believing in it in the beginning. Also, I mean, so the data sources, especially when I think from the perspective of we being like a memory organization, preserving the lineage and the history of the data is important to assert the change that has happened without knowing the past. You cannot say something has changed, right? And it is sad to see some of the government funded organizations that were hosting climate related data. I forgot the name of that site right now. They basically announced that they won't be maintaining the data anymore basically and so we need to jump in and kind of archive as much as we can and make it available in a meaningful way. But there is another perspective to it. While we are archiving, we need to be mindful of not causing more harm into the system. After all, archiving is nothing, but it's another digital or computational infrastructure that we have. And if we tomorrow start like, hey, let's index the whole web as best as we can. Maybe that will come up with a hefty carbon footprint, right? So how to do it efficiently and smartly without harming the system is, I think, it's a process that you are still learning. So with that, I think we run out of time and thank you so much for your questions. Thanks, Quinn, for joining remotely. It's a pleasure. Thank you so much. Thank you for having me. All right, bye-bye.