 You basically have to do the the work of the seeding and sorting in Chrome and so I'll just start from the very beginning. Okay, so if you start in Chrome you can go to the edgy website which is in ByroDataGov.org and we have all of our archiving primers available if you click on the archiving tab and then if you click on the agency archiving primers list and so this is where we have all of the primers that have been written and sort of worked on for edgy. We have two kinds of primers. The first are main agency primers and these are based on an entire department and so what this is is sort of an overview of the department, the transition team if it was pre-inauguration or now sort of the political nominees and then also a risk assessment which is sort of just comments made by the administration or sort of targeted regulations or things like that. What you actually use for the web archiving aspect are the sub agency primers which are listed sort of below and they're for specific offices in the department and so if we were going to be working with this department of energy one we can take a look at for example the Office of Science subprimer and so you click on it and it opens up the school doc and there's some background but the main important part is essentially this list of starting points and what this is is a list of the important parts of the website that we've essentially pulled out that are good starting points for people to start exploring the web page and so what a subprimers purpose really is is to be able to act as a guide for people that show up to departments or show up to data archiving events that don't have prior knowledge of how the departments are laid out and how to find data there and so you essentially assign one starting point to one volunteer and they sort of explore the web around there and so if I were coming to event and this were the the subprimer that I would be working with I could be assigned sort of this starting point seven which is this base the basic energy science research programs and so what I would do is I would click on this link and this brings me to a place in the Office of Sciences website that I can now explore and so the numbers that I need to know so the Chrome extension uses some numbers for metadata essentially or labeling and so up here you can see the Chrome extension it's called the nomination tool and I can show you how to download that and you can see that there are sort of the title and the URL are auto filled there's a space for your name your email which needs to be in the form of an email but doesn't actually need to be an email and then the event name and then these codes and so the agency subagency organization and suborganization come from the agency office code so this is a unique for a number basically code that identifies this particular office so what you do here is you would say that it's 2310 and that means that every link that you submit like this will be tagged back to the Office of Science then the subprimer number is the number in the numbered list here of the starting point that you're working on so in this case it would be seven and this just helps to further narrow down where in the website this came from and so if you wanted to submit this page what you would do is you would go 2310 and then seven and then you would take a look at whether or not this page was crawlable or not and we can talk about like the distinction there this looks fairly crawlable and then you would hit submit and then it would say you are all submitted thanks and then you're all set and then what you would do is you would basically click through the website by clicking on these links and sort of seeing where you can continue to explore the website and so what we're looking for for seeds are not PDFs but we're looking for a place in the site map of the of this website where there are sort of places to go so what a web crawler does is it clicks on all of the links on one page and opens those and then it clicks on all of the links on those pages and opens those and then it does that for three levels in the case of the end of term web crawler and so that means that if you give it the start of the site tree essentially it won't make its way all the way down but we're trying to give it these intermediate points so it can still fully explore the site map and so when you're looking through the web page you're looking for sort of not necessarily PDFs to seed but you're looking for web pages that for example here will bring you to pages with more links you can click into here and this is sort of their pages with more links here so this would be a good seed and so you're approximately trying to hit it every two to three levels so that you can sort of accurately cover this site map and so then the difference between crawlable and uncrawlable is that when we're doing web archiving we label pages as either crawlable which means that they get sent to the end of term harvest for for their web crawl or we label them as uncrawlable which means that for whatever reason the web crawler would not be able to work on them and there are sort of four main categories for that and I can go over that in a second but these are then sent to the data archiving web workflow and so they're put into the archivers app and then the data is sort of harvested that way and so in the nomination tool you can see that there are sort of these four toggle switches that you can look at and these are the four sort of kinds of uncrawlable things that we're looking at and so what we can what they're what we have is basically this FAQ that outlet I can email you that basically outlines what are the different kinds of uncrawlables and it gives you an example of each of them so for example a file transfer protocol looks pretty much like this and this is something that we we send to sort of the data archiving pipeline if it's a visualization or an interactive if there are just a lot of files so that you can easily pull this down with a script or if it's any sort of queryable database these are all things that are uncrawlable although FTPs are crawlable but we send them to the uncrawlable pipelines that they can be harvested properly and so then you can mark this in the Chrome extension and submit it in the same way and so if you don't toggle any of these switches it's automatically marked as crawlable and if you do toggle any of the switches it's marked as uncrawlable and so you basically work your way through the website like that and the way to download the nomination tool is that on the top right hand of every page of every sub primer there's this document linked to the agency office code which brings you to a place to actually download the Chrome extension and then also a quick how to recap if people get confused or aren't sure what to do next. So does that make sense?