 Okay, thank you. Thank you very much Gwen. So let's start this session on the main services of OpenAir Provide. So session dedicated to the services that we have available and all the interactions we have with our content providers. Repositories, open access journals, Cree systems, data repositories, data archives. So those that are managing these repositories, these different systems that are providing content to OpenAir. So the main objective is in fact to discuss some of the services available to discuss with you and to present you the latest updates on our graph, on our Provide dashboard, on our interoperability guidelines to highlight some use cases from the graph and the way that content providers can consume the data from our graph. And the way that users and users, repository managers are using our Provide functionalities services and the way that the community is following our interoperability guidelines and actively participating in contributing to the development of the OpenAir guidelines. So it's important to say that we present and we discuss all the services around Provide in a completely open and transparent way. So we are trying to improve the way that we engage with our users and that we receive feedback. So just before we start, I would like to invite you all to those that are not aware of these community calls that you join our community calls. So every month, every first Wednesday of the month, we have community calls. We realize that this is the right way to go with to welcome your feedback to have a more participatory way of welcome your ideas about our services. Be aware of these community calls. We try to have them all every month. Open for discussions. We always have a specific topic to be discussed, but we have plenty of time to discuss other issues to clarify some doubts about the services about infrastructure, etc. So be aware of that. You have here in front of you the specific dates of the coming calls and the first Wednesday of each month. Be aware it's always at half past to Central European time. And we hope that you can join in the coming and also be aware that we have monthly newsletter dedicated to content providers. So if you are an open air content provider, or if you are someone interested in the provide services, guidelines, broker events, user statistics, the validator, etc. So subscribe this newsletter because we will always try to send three or four news items every newsletter. And this newsletter is published every month just before the community call. So by the first Monday or Tuesday of the month expecting your mailbox, the open air content providers newsletter. So, so I have already give these two invitations for our newsletter and for the community call so let's go to our session. Thank you for very much for the speakers that we have today. In fact, our initial idea was to have more use cases. I will highlight some of the use cases that in fact they were already presented in several community calls, but we have a mix of use cases and people from open air presenting the services. So we have seven speakers organized in three blocks. For each of these blocks, we have a specific moment for questions and answers. Okay, the first block is related with our research graph and the way that content providers can use and consume the data from our graph and and the way that we are generating added value services from our research graph. So I'll do what sorry from CNR is T from the open air technical team will present when a research graph in the broker service, the power of the open air research graph into in practice, it's important to say that now we have in production, a relevant resource so more than 100 million records available in our in production from from our research graph so it's important to have this idea and cloud you will present this, the services the added value services that we can build relying on this graph, and then a specific use case from my reference from the Latin America network of repositories by long time. So thank you, Claudio and long time for this. We will have it in the first part. And for this first part we will have questions and answers for let's say that we have often our for each block and for each block we have questions and answers the second one is about the specific functionalities and services of provide I will do a presentation of the different dashboard services and then we will have a kind of zoom for a specific service that is being updated and and and and improve it in the rest of months is the user statistics the users count service that Dimitri Spiracus from Athena research center that is part of our technical team also responsible to manage the service will detail and then the third block and for this we will have also questions and answers for for the third block. It's dedicated to the to our guidelines. And then we will have our colleague from Villa felt University Andreas will detail the recent updates regarding the, the, the different, the set of guidelines that we have available for our community, and then to use cases from the literature, the publications guidelines and we will have our colleagues from from Canada. Canada research open air compliant Canadian aggregator perspective from a national aggregator compatible with the open air guidelines and again, the drug rack from Charles University from Czech Republic will give us some important novelties about the crease the crease guidelines. So, let's start first with the the open air research graph cloud you the floor the floor is yours be aware that for it for each of this block we have questions and answers. Of course, we can also have it at the end but if you have something relevant to ask about the broker the research graph and the specific use case that long time will present so we will have time at the end of this first block. So, I will stop my screen and give the floor to Claudio. All right. Thank you, Pedro. Let me share my screen. Okay, can you see my screen. Yes, I don't see your camera but okay. Oh, right. Here again. Hello. I like to see beautiful faces. Thank you for the introduction Pedro. With this brief presentation I'm going to give an overview of what the open air research graph is and how in open air we are exploiting it to build another value service like the broker service on top of the information in the graph. Let's start by defining what we mean, essentially by graph. We define it as an open metadata research graph of interlinked scientific products with access right information linked to funding information and research communities. This is the definition that we did in open air to the graph of information that is aggregated by open air. Generally speaking, a graph is a model to represent information. You can think about the Facebook graph with people as dots and relationship between people like friendship as the edges connecting the dots. We apply the same concept to the scientific knowledge. In order to have the possibility to navigate from the dots in this graph. In fact, as you can see in the picture here research products can are associated to the research projects that funded them. Instead link to the funder entity. So, following this relationship you can get a given perspective about them for example the expenditure for scientific production, let's say. Oops. Sorry. I changed. Present. Present view. Here we are again. Okay. So, more information about the open air graph is now available at the newly launched website graph dot open air dot you so you want to know more about what I'm telling you. In this moment, you can find more information there as a lot of resources have been introduced there to better describe what the open air research graph is. So all begins with a quite acquisition of metadata about scientific products. So, as today, we open air has more than 12 K sources spread around around all the world with interaction with many many actors that you might know and initiatives scholarly communication initiatives that are available out there. So, the kind of information acquired in the graph is not limited only to articles but spans also to, let's say new type of products like software that are emerging in the movement of open science and evaluation of science itself. Data sets are becoming more and more prominent, just like research software so this is why you have also GitHub here in this slide. Another very important concept to keep in mind on the way open air builds this research graph is the role of entity registries, which we define as a particular type of data sources that offer authoritative list of entities. Among these we can mention registries for researchers like orchid that deliver non ambiguous identity for in this case, the person that is a researcher. Then we have registries for organizations like greed.ac and of course, Cordis our main provider for and the first provider actually for project information and the relative funding stream. Then we have registries for projects, again Cordis plus several funders scattered around Europe but not only to limited to Europe. And last but not least registries for data sources like OpenDoor, R3Data, DOIJ for the journalist and the publishers. So the concept is that from this particular type of data sources, OpenAir acquires information that is non ambiguously identified and the identification of the identity as we can shall see in next slides is going to be a central point in the construction of the graph. So the story begins with the aggregation of this metadata information from a different from different sets of data source typologies with I can say the most common one are registries delivering metadata information via OAPMH protocol. My colleague Andreas later is going to give you more information about this particular type of data collection. And thanks to the role of the OpenAir guidelines, I would say the integration activity for this type of sources allows to have now thousands actually of sources of this typology because they follow the rules expressed in the OpenAir guidelines. And then in the slide, there is another block below where other type of sources like Orchid, I'm Paywall, Crossref and Microsoft Academic, not following the OpenAir guidelines are instead aggregated differently, but allowing to build a quite comprehensive set of scholarly communication data. You can see the numbers here, far beyond the 100 million records are ingested in the system, forming what we call the raw OpenAir research graph, raw because it's not yet processed or enriched. Then the supply chain that is applied to the graph assumes to look for multiple instances of the same research product, publications, software, data sets and other kind of research products. For example, the preprint could be acquired by a given repository and the postprint by a journal. So the goal for this process is to identify them as the same object because we don't want to count them twice in the statistics and we don't want to see it appearing more than once when a user issues a search on, for example, the Explorer portal. Then thanks to full text mining algorithms and set of different kinds of enrichments that we perform, added value, let's say is added to the graph. This added value lies in giving more uniformity, trying to map different terms that can be expressed in different ways. For example, to state the language of which publication is written. Different repositories might express this concept in different ways. So an important role here is to apply mapping, transformation, thanks to control vocabularies. Finally, this step feels the final representation of the graph that is finally pushed to the back ends serving our public services like monitor, explore, connect and the API through develop. As you can see, the amount of services that can be built on top of the graph is quite important. The graph in this sense plays really a central role in many of the things we do in OpenAir. So from the next slide on we are focusing only on the broker service that is part of the OpenAir Provide dashboard. So the main concept behind the broker is to give back this added value, this uniformity of information and enrichment to repositories that provided this information in the first place. So the added value built along the OpenAir supply chain is going to be given back to the regional repositories to enrich their regional collections. So the purpose is to allow repository managers on the Provide dashboard to evaluate the quality of the kind of enrichments that OpenAir can build for them and subscribe if they are interested. So this flow of actions is well-descipted in this slide where thanks to algorithms that I'm going to illustrate briefly in the next slide identify the events that are presented to users up to the top 100 events as a preview to the user. Then a user can create a subscription once he found that given type of enrichment suites some lack of information in their collection. Then periodically subscriptions are matched against the events that can be produced for the graph to form what we call notifications. These notifications represent the actual enrichments that the broker can deliver back to the repository managers. And this can be explored in two ways on one side on the content provider dashboard and can be for the moment we are testing the realization of the public API for the broker in the beta environment. So to consume notification programmatically so to enable automatic integration of these added value information back in into the repository collections. So how does the algorithm to identify these added value works. The algorithm can be well exemplified by this example the concept lies on the fact that we run the application. So the algorithm is based on the analysis of the groups of duplicates objects so the procedure can be well exemplified by an example. Let's say that record a coming from repository are the a is a duplicate of a record be from another repository. Let's assume that a provides an abstract while record be does not has an abstract, then the algorithm can produce an event potentially interesting for the other repository. And the same criteria can be basically applied for other meta data field might they be alternatively provided by other repositories or inferred by the mining algorithms implemented in the open supply chain. It's important to note that since events are produced by algorithms that take decision based on non authoritative information. It's crucial that for transparency this level of uncertainty is made available to the end users. This is why in the content provider dashboard when exploring the events. There are some sliders that allow users to filter also by a concept of trust that maps what we can synthesize as the confidence level. It can be produced by the mining algorithms or the. Outuitiveness level being higher value more near to one. When the information is provided by a repository. Enrichments events are presented in content provider dashboard are modeled inside the broker as two main categories. Enrichments of more information that was recognized as already been available. In, in a repository collections like here's another persistent identifier for this publication or missing information. This publication did not have any persistent ID. So here I do I. There are different categories of enrichments we call them topics that touches upon different meta data fields like the abstract publication date links to projects subject classifications obeying to different known taxonomies. Open access versions so URLs links to the open access version of a scientific product persistent IDs or kids and we are still evaluating the quality levels of other type of events like links between. Literature and data sets links to software and links to other publications. Just to show some numbers produced some time ago these were. The amount of events built for the different kind of topics for. Yeah, back then they were for something like 700 repositories that are available in production. And this concludes my presentation. Okay. Thank you very much. So you can put you can ask questions here in the Q&A or in the chat if you want. But let's move to the to the second presentation of this first block to have you hear them. Use case and then you can discuss a bit more about this first part of this value of the opener research graph for content providers. Thank you. Thank you for the opportunity to present. I want to briefly present what we are doing collaboration with open air to consume the services that they are generated. And as you know open and references and aggregator is not a repository. And it's a content provider for for for open air. We open our hardest more than two million records of Latin American repositories to reference. This use case is different from the ones from from most of the of the of the cases of open air because we are aggregate aggregator of aggregators because we have this actually the national notes of our references so First, I want to talk about what are the ways that we are developing to to consume the service that open air is provided. The collection of a reference is part of the of the of the records that that feeds the graph. So we have a lot of of of information that in the graph that that this is covered by the process of the graph. So, we develop a repository dashboards for national notes and our references repositories because they are not a part of the that part of the of the open air provide. We developed a repository dashboard in order to present services of our reference and also consume the services provided by open air. This is the way that we that we think that is better to to to connect the two initiatives. Another integration that we are experimenting is to consume the open air graph a IP from our search site from from our provide site. In order to present additional information about the records of reference to the final users. So, we are going to, this is a beta version is a prototype. But we are going to present additional information provided by the open air to the old the to the graph for the graph a IP directly to the final users. So they will see the data, the metadata from their regional repositories, but also how that metadata was connected and connected to other providers by open air graph. And one thing that one also want to show is how the impact of the of the of the graph. In the national collection in the open air explore, because we did some work with open air team to improve the way that the reference collection is displayed in the in explore. Before this work, la referencia collection was always perceived as as one thing without this wasn't between repositories after this after this this work. Now, you can know that this this part la referencia is the only organization mentioned here, but now after the improvement of the opener explore interface but also think by the work that we don't know about providing providing some extra information in the records and individual repositories are being identified and presented in the opener explore. As I said, la referencia is a rational repository network, we harvest national knows that harvest repositories. So, before this work, open air harvested a processor reference as a monolithic collection. So we did some work with the with the opener team in order to improve the way that the information was traced. So we have been in the regional repository in Latin America to register in in open door. That is the way that open air use to one of the way open your use to to identify and you know repositories. So we improved the data of the of the records to to put the open door ID in each record in each with the data record, then open our team work in the pipeline and harvesting process to to discover this information and put it in the process. So now, as a result, individual repositories are identified and displayed and use it as, as individual sources in the in the graph. So we think that it was a very important step in the collaboration between between our reference here and opener. And the next step is to better integrate the broker AP in the, in the reference repository dashboard in order to deliver the events that global show it directly to to to to Latin America repositories behind our reference. In general, the the opener at my project that is going to end our reference he picked and center. We are exploring ways to, to, to use the outcomes of the graph to to integrate that information into our software. We are working in in metadata arrangement and link it with Chris and national system, our region so we think that the opportunity that gives the graph and the service that open area, open area is developing is important for the other regions. So we need to find a ways to collaborate and integrate information, not only as a provider, but also as a users, or even to deliver better service to the end users. So thank you. That's all from my part. Thank you very much. If you want to join also probably feel free to join this to connect your, your camera and audio. Let me check if there is questions here. Thank you very much. Catherine is asking him. Thanks for the great and very clear presentation about what it is open a graph. Very helpful. As you know, it is quite an effort for requisitors to adopt the guidelines version four and different visitors don't adopt the guidelines they content. Is there still be included in the open air graph via other data providers. Can you explain what are the specific benefits for the visitors to adopt the open air guidelines. Well, I believe the idea is to reduce the effort also for others to integrate the information. Unless we want to let others do become the standard the factor standard. I think this is the main drawback, let's say adopting the open air guideline will surely provide another value for us, the open meaning the open air team, because the effort for us to integrate that information will be lower. And agreeing on an open standard for method exchange is another value, not only for few but for a broader audience that say so it's for for the benefit of many that is good to adopt. Open air guidelines. And of course, and then for sure, Kathleen knows as so well as as as we, because contributing to being compatible with the the open air guidelines will also provide the possibility to align at other levels like also at the national level, and also to some kind of national infrastructure that there are in place, and align also with other kinds of services and benefit from them from that. But, but for sure, maybe we can highlight some examples that aligning with the, for example, the last version of the of the literature publications guidelines will provide us information. That we don't have from my other means like things like author ideas, this kind of enrich the content that is quite important for the open air infrastructure and also to open air to offer edit value services on top of this better quality, the better method of the quality. Let's say. I jump in. Sure. I was about to ask some from your team could reply. Thank you for speaking. And good thing is to have one of the fear principles. So, if the metadata are has a high quality like with excess rights and licenses and so on. So I would like to find and to reuse by other researchers. So I would like to give some objectives in the my talk later. So can you please share the, the, the one of your last slides. Stephie is asking here please display the slide and the steps for setting up using the, the open air graph with code, etc. I think it's this one present. Was it this one. The different phases. I think so this is the one before just the next steps is the slide before the next steps could you change. I don't have a slide next steps. No, no. Near near the end. Yes. This one with a new type of events. Let's wait for a reaction. As I don't have your slides yet because I tried to. I don't know I cannot help. But let's see. Could you go to move just to the, to the one of the last slides. Well, the second last. This is the one before the thank you. Yes, maybe you can ask the question. And we should have shared the slides before we will share the link here in the, in the shot quickly. If you have any other questions, please ask. I see some comments here in the stuffy the one after this is the, the last slide. So maybe can you ask the question. So I'm not sure if you can you. So after this slide, it's the last one. No. Indeed. Yes. So maybe we can address your question selfie after. Not sure if what you want just ask here what you want to know better. Maybe we can explain with the code, et cetera, with the code slide with the code. No, I'm not sure. So here we have the different events that we have that we have. Okay, we will, we will share the slides in the coming in the coming meetings here in the in the link and then we can we can address. Once again, for sure we will not avoid to, to answer your, your question. Okay, thank you very much. So let's, let's, let's proceed with this discussion. And we will check once again the question and answers and let's let's now just move to the second block. About the provide where we, I will I like some of the services and then my colleague Dimitri will talk about the users count service. Very briefly, the, what we have in this dashboard and I like some, some use cases recently presented in some of the community calls. Normally I like to, to associate this image to, to the, to the dashboard to the provide dashboard I already used a different kinds of images and different kinds of graphics. But I always like to use this image because this is something that of course, it can be also an image where we are playing with different services and functionalities but that this is what we really want to have. As a community we are pushing ourselves to, to have the different kinds of services contribute to contributing to this global open research community becoming more powerful. And this is something that provides service can, can, can, can in fact provide via the process of validation via a process of enrichment. The, the open air provide dashboard is in fact this services, this service where all the kind of content providers can interact with with open air and become part of this global community but also we want to have to provide as the gateway to the European Open Science Cloud and in order for some of these content providers actively participate and be part of the European Open Science Cloud. They need to be part of open air infrastructure they need to become content providers of open air infrastructure and contribute by this mean to the and be part of to contribute and to be part of the European Open Science Cloud. And components different functionalities in provide. So we have the, an area to validate to register to, to, to check metrics to measure and to enrich. The validate is the services related with the, the validator, the, the metadata validator that we have to check compliance against our guidelines, a process of registration that we align with other alternative directories, like with the data and the user statistics, where we check and we provide information about metrics from the content providers and the broker events. And the broker content that is that cloud you have already presented available in a part of in a component related with the content enrichment. It's important to say that the current the version that we have currently available in production is the second version of this of this service. We, we did an update last May. We prepared during several months is relying on relevant contributions from the community and the first version was presented and delivered in the, in October, September, October 2018. We did slightly improvements, but then we delivered a new version with a completely different layout in, in, in April, May, this year. So, the, just some highlights about the, the main functionalities in services available, the one related with the validator is in fact, where people can run compliance test against our the different guidelines that we have available for repositories for journals for data repositories and for create systems. And this is a relevant service in order to, to people to check the quality of the metadata to to receive kind of report with the feedback and to check where they need to improve. And this is important to do when we want to register for the first time but it's also important to do when we want to improve the, the version, the level of compatibility that we have in our data source. So we can also test and run different tests in the in our validator. We have this available for this main kind of types of data sources registered is also associated to the validator of course because when someone want to register. We also run validations but the register the register and functionalities just for those that are for the first time to be part of open air into and into register they are their content provider in our infrastructure. There are automatic service available for publication repositories for data archives for journals and for aggregators. Of course for literature and for that the repositories we are relying on a, we have, we are relying on an authoritative directory. We have a new structure repositories and every three data for that repository so I knew that the source need to be in this directory and then they can register in open air and and comply with our guidelines. We expect we are preparing new automatic process of registrations also for Chris systems that it should be already available it's not yet. The scope of the guidelines is already available but the automatic process of registration is not available yet but for those that have a Chris system and that want to be part of open air of course they can contact that contact us by the help desk or by the other means that we have and we will initiate the process of registration what we don't have available is just the automatic process inside the provide and we hope to have different types of repositories and systems also available soon as we also have we propose also guidelines for software registries software repositories and of course we deal with different realities we have kind of repositories that that are kind of hybrid repositories where the where we have all types of contents in the repositories and we need to find good ways to to to process the registration in open air so soon we will have it, but it's important to say also that under this umbrella of the the register in open air, it's possible also to update of course after you are part of the infrastructure, you can benefit from from all the provide service. You have also a possibility to update information basic information from the, from the, from the data source from repository, or, for example, update the way IP image interfaces, or some kind of information like logos or descriptions from your content provider. This is also available under this register umbrella. The provide service, of course, within this validator and registration service, of course, is related with our policies, be aware that our policies in our guidelines are available in our main portal but you can also access them via the open air portal so please check the guidelines and check the content acquisition in order to understand all the useful and the needed information for you to register and to comply. Then the broker service already well presented by cloud you it's available in a area about enrichment content enrichment so we are delivering metadata events that can be useful for the repository manager for the end user to use it in your own content provider so this is available in the content storage area where you will have information metadata events to improve your metadata and we have different kinds of, of, of events and we will have different use cases that I will present in a minute that people using the broker events. The fourth area of functionalities and services available in the dashboard there is the one related with the user statistics where we can check views and downloads and the numbers from your specific repository in open air and you can, and you can get the service and benefit from this aggregated environment so like a kind of metrics hub that we will provide to you. This will detail this in some, in some minutes. So, to finish, just some highlights. So, the service is available in provide.openair.eu and of course you can access and know everything about the service connecting to this page. You can also log in and have some more details about the, the provide of course you only if you have a content provider, a data source that you manage. You have access to this different functionalities. Okay, you can log in, you will see useful information about the service but you don't see information about your data source only if you are the manager of your service. If you think that you are the manager of a service and if you don't see that information, just contact us because there's something wrong with your login or you are accessing the provide with a different login that in fact is not the login associated to the admin rights of your content provider. It's important to say this is something that we, I think we have put a lot of effort. For sure, it's not perfect, but we have put a lot of effort that this current version of the provider, the version that we have released in early May this year is a result of a participatory process, several interactions. So we did them three, four workshops where we did demos and we collect feedback from from end users, some also webinars. And after we released this version, in fact, we did some more user tests to check some of the functionalities, we are quite happy with several components of this but not completely happy with one or two things that should be improved, but this is life, life is not perfect, and also the service is not perfect, but we are quite happy with what we have in the way that we have improved in this participatory. Here I have some print screens from some of the user tests that we did with several repositories from different countries here are represented from Italy, Spain, and Turkey, for example, and we were quite happy with the results. Because I think the dashboard itself now it's much more closer to the user needs. It's much more a dashboard environment and we and for that we are, we are quite happy. And just some highlights. So here you see the different components where you see the register and the validator and the notifications in the left side of the screen to benefit from the registration service and the validator notifications are related with the broker. So if you subscribe to specific metadata events you will receive monthly in a monthly basis, the notifications. Then you can update here in the, in the center of the screen you have update aggregation history enrichment and user statistics are the tabs where you can interact with the different components of this of this dashboard. So the aggregation history is something quite important. We are not completely happy with the way that we are communicating with the end users but, but for sure this is something that users are requested a lot in order to understand. When was the last time that open air have aggregated content from their repository. From which date the they have the content available in production in open air, what when was the last time that the content was indexed, etc. And everything is available in this aggregation history. And this is something that provides more transparency to the, to the open air infrastructure that is, it's complex and it's becoming more and more complex due to the expansion of the graph but we need to communicate straightforward to the, to the, to those that are contributing to the infrastructure. And then those that are contributing can benefit in this tab of the enrichment and this tab of the user statistics, where you have the metadata enrichment events and the user statistics. And then you have the numbers where you can benefit from these two services to finalize some examples. We try to have different calls the community calls that I have already invited you in the beginning of this session to participate. We try to have a different to present different use cases. I think we have, we have already several but we have been five highlights at least. We have two repositories from Italy, Serbia, Portugal, journals and the, and also Spain. Two highlights. We have a repository from the university from Trieste, that is using their display screens repository to improve their interoperability quality using our validator, but they are improving a lot the method that the quality using the broker events. And the, the enrichments related with open access versions that we are suggesting to them. They are using the PID events that we are suggesting to them. They are using the, the, the, for example, the abstracts that we are sending them. So these are some examples that our colleague Jordan from University of Trieste presented in one of the community calls we did this year. My colleague Ricardo here from the University of Minho that is the manager of this important repository in Portugal is using a lot the broker service to improve the metadata when it comes to related with them links to projects. So, all the links related with the Portuguese funder and the European frameworks like Horizon 2020 and FP7 are improved in our metadata as a result from the metadata enrichment events that we are sending back to Ricardo as a repository manager of this. And the same is happening in several Portuguese repositories. Also, when it comes into registered validation enrichment. The network of repositories in Serbia is all those repositories at all these 16 repositories that from Serbia are using the, the, the registration and the metadata and the validator service. It was quite helpful for their quality of metadata, and also for the improvement for the enrichment of their records as, as, as our colleagues from Serbia have presented, and then it's useful also for journals for open access journals from different countries, also to become part of open air and to become and to provide more visibility and and for them to, to improve a bit they are metadata, metadata specifications from the journals. So, different use cases that are being presented and and then say in some countries that are in an interesting way to participate there are lots of developments and recent initiatives where they invite people from open air to to better represent the provide and as an example of this recent initiatives at the national level is this this activity that our colleagues in Spain from the network of of libraries in Spain are organizing a session on the 29th of October. In fact, dedicated only to provide it's a workshop from the repositories of this network of libraries, where they call it three experiences if I translate directly from Spanish to English. There are three use cases using the the open air provide service, and they will have the University of Polytechnical from Valencia to present a broker use case, and a provider from University that Carlos Tercero I'm not sure if she's here with us, and they will present the way that they are using uses counts. And Tony Prieto from Polytechnical from Catalonia they will present the way that they are using the the API is in the in the graph to enrich their content and for example to check duplicates in their repositories based on the on our on our bulk content. So these are three different use cases that in fact you will be presented in the community, a national effort that this colleagues from Spain are doing we are happy to support if you want to do similar things in your countries. I think these are great examples that I would like to share and thank you for those colleagues from Spain to organize this. It's all from my side. I'm happy to answer questions I will check later here in the chat but I invite now the meetings to do the presentation of one of the services that we have in provide that is the uses count counts service. So the meetings, you can share your screen. Sorry, I'm trying to share my screen. So let me check if I can help so I can open the slides for you the meet the meetings you can. If you cannot manage to present sometimes there are some issues. I can, I can open the slides. Okay, can, can the meeting. Let me check. Let's give 10 seconds more if you if you cannot do it, I will open the slides and share the screen with you. So, the meeting is not here. I see I see him removed. Okay, now the now the microphone is there, Dimitris. Can you try to say something. Yeah, sorry, sorry about that. Sorry about that. I had to share my screen to restart my PC to zoom up to share my screen. So, can you share my screen now? Yes, yes, yes, you can. Can you see it? Yes, perfect. Okay, okay. Thank you. So you can start the presentation you can present yes present mode. Okay. Okay, thank you. Thank you, better and sorry for inconvenience. And I will try to give a brief overview of the opener. Just as count service, which is a part of the services offered by provide and discussed previously by Pedro. In this overview will try to be given introduction the service, what this service is about who is it for how can you benefit by using the service. And how can you participate and join the service and most importantly, what results you can get by using the service and how can you get these results. So what is the user account service in the inside the open as infrastructure. In the open. In this infrastructure, we have the con providers, the institutional depository the aggregators the, the software depositories which Claudio mentioned in his presentation. And that content to open Earth is content is a mind harvested and the obligated in order to produce the open areas, a reset graph, which includes all the research products for example publication software is that the usage count service links the usage events to these projects and by linking we simply we mean simply counting these usage events events as they occur in data in data sources. The usage events are views and downloads for its research project and the usage events are collected are aggregated are analyzed and we produced and finally produced statistics. In order to deliver standard activity reports about research usage. So, this is more or less the usage counts service inside the opener infrastructure. And how do we do it. And this is, we have followed two approaches to different workflows for your workflows for usage counts. The first workflow, a workflow is named the push. Which is used to collect usage events views and download directly from data sources with collect them we process them using some standard rules which are provided by the counter code of practice directive. And these rules allows us to remove. For example, duplicate records or non legitimate traffic, like robots or spiders, and we collect this events to our user statistics database. And the second approach is named the pool approach, where we collect counter reports from national aggregators for example, I was UK. We store these reports in our user statistics database and we publish them either in open interfaces like explore or provide, or we can deploy them by the API endpoint which follows the sushi light protocol. So what are the main features of this service the usage count service. First, we, as we mentioned, we track using downloads by the push approach or we collect counter reports using the pull up approach. We support IP anonymization in order to offer a kind of data privacy. The important thing is that we exploit metadata duplication that enables the accumulation of using downloads for same documents. What practically mean is that if we have an item which has a with which for which we have collected usage activity in many repositories, then we can have a usage events user statistics from different repositories. The users count service supports the counter code of practice. It's compatible with counter code of practice release for and we will develop in the future the combat ability to counter code of practice at least five. And this combat ability allows to produce standard base user statistics, and most importantly, it enables the compiler to compare our statistics are used statistics with statistics from other data sources. So to recap the usage counts. What is the usage count service since everything counts in large amounts. So, a usage count is a measure of scholar impact. It provides indicators that complements other biometric indicators like citations in order to offer comprehensive and more recent view of the impact of academic sources, who can benefit from the service who are the stakeholders of the offers institution of a science platform funders etc can use the service can exploit users count service and answer questions like which founder has the biggest engagement in Europe based on the usage activity or providing the evolution of the popular to the publication of a project within the last five years. And we can use the service to build inference or prediction model for topics based on the users activity. We create user communities, a group of people with with with share common preferences and make recommendations. And of course we are provide as mentioned before you via the counter code of practice directive required standardization. How can you use the service can can join the service of as Pedro mentioned you can join the service by registration of the providers dashboard. From there you can find all the required information and the steps that are needed to be performed in order to join the service for example you have to download the software. You have to provide software for your depository you have to configure the tracking code and deploy the software and from outside the opener side we will validate the installation of the software and we will inform you accordingly in order to be able to you to view your user statistics. For now we offer software for for this space from version four to version six as an as a patch for this versions. We offer a plugin for version three for reprints and most importantly we offer a Python script for all other key for all other platforms, including of course the space and the prince that can be used independently of the repository platform. So we can install with our help, if it's needed to send us as events to open a usage count service. You can see the sample numbers you can see the evolution of open air users count service from 2016 to September 2020 you can see the user now loads that have been collected and aggregated. And you can also have an indication indicator of the open a dialog downloads for the various for funders. In Europe. You can see how this information, how the information that we collect can be exploited for by this particular stakeholder. In this slide in a Pedro's presentation, you can. This is the usage counts. This counts in action and the user statistics which are published in a particular repository. You can see the number of using the number of downloads that have been collected across years, and you can also use this interface, this interface from the provide to retrieve more customized reports, for example, article reports, during a certain period of time. And you can also download the users reports using our sushi light and points, we support counter a kind of practice release for compatible reports, and we offer article reports book reports repository reports for general reports, apart from, and of course item reports. And this is an example of the report that you are that you could get using our endpoint. And this is in the sushi light format that are mentioned before, and only left hand is the repository report, and on the right side is the item report. Recently we have deployed a new kid in town it's our usage counts open portal, which provides information about the usage count service includes various resources includes a set of analytics that are involved every day. And we have a contact form where we can ask anything you want about the service, and we have information about the architecture and structure of, of the service, you can go there by using the usage counts that open air dot you link. In this portal, you can see a number of analytics like the, the, the current status of the results that we have that the user data that we have collected we have 200 repositories 100 million views and 300, around 300 million downloads, and you can also search for a particular country to use the statistics. The usage activity for a particular county for example for Portugal we can see that we have 35 repositories 27 million views and 40 million downloads, all this information is provided in this new portal, the usage counts portal, and as mentioned, we will add more and more analytics there. In the future, we will update our infrastructure, we will try to build the service in a more robust infrastructure, and we will offer the service in a new backend that will be that will have a better performance, and will be more stable. This is the usage counts service team, apart from me is Andreas and Joachim from University of Billfield, and my colleagues from Athena Research Center, Adonis and Lebesis and Spiros, and Japan. So this is all from me. Thank you. And if you have any questions, be available to answer. Yes, Dimitri. So, before we move to the last topic of this session, feel free to ask questions. I saw already some questions and in fact provided already some answers. We can also answer directly in the text if you have questions later to put to Dimitris. I would like to highlight one of the questions that that in fact is being also discussed here in the chat, which is good, which is great long time also reacted to Ellie from the Netherlands. And regarding the, the, the, the Nazis so it's important to and I would like just to take 15 seconds to that so it's important to say that when open air collects content from national aggregators from, for example, European national aggregators. And we have the content from those from a specific repository from a specific country in open air infrastructure via an aggregator. Unfortunately, currently, we don't have a way to provide an individual access to the repository managers in the dashboard. Okay, this is in the pipeline. We had a, this is one of our priorities in provide just to make sure. Unfortunately, the priorities. So, we cannot achieve sometimes what we have on the top three priorities because we have the priority that don't allow us to progress in other things so the priority was the graph expansion. It's solved now. Now we can move to three or other priorities and I think we can move to this one. The way that we solve the technical limitation is like long time ago said in the shot so it's is similar to the approach that we did for la referencia repositories, but that we can manage that. Now we know what we need to improve is in the also in the in the front end, and to create some, some to consume some information from the data source manager in internally in the back end of open air. So, if cloud you have some additional information about that, you can, you can say it, if not, this is a kind of, it's a, it's a commitment from the technical team, it's a commitment, it's a personal commitment that in order that we will have it in the coming months before the end of the project for sure, but in the coming in the coming two or three months is for sure we will have this available to support not only nurses but other national aggregators okay repositories. I know that the Dutch repositories are asking for this. Since the first time that we present the dashboard, every time that we present the dashboard. Dutch repository managers ask for this. Okay, I am aware of that. So, thank you for your question and this was a possibility for me to to clarify this. If may, if may I, I would stress what Lautaro already replied in the chat. Perhaps a path to proceed is to exploit what we did for the reference essentially asking aggregators to include the open door identifier of the original repositories being aggregated by an aggregator so that on the opener side we can recognize which the deposit the original repository is in this way we can, for example, generate events directly for individual repositories in the broker. Yeah, true. This is the important information for all. Sorry, Pedro. Just one, as Claudio said, if it's useful what we did for our reference here we can share the way that we did because opener is already prepared to process those additional metadata records so if it's useful for other content providers we will be happy to share the process. Yeah, thank you very much. This is quite important for other aggregators, and we will work with you to put it in the support materials that we have in provide, but as we did already some work specifically for Nazis. So I know that some we did some progress. So this is not the, this is not the case. This is the Nazis is waiting for analysis repositories are waiting for the other part. So the way that we can, in fact, put it in production in the in providing order for a repository manager have access. They have already done some work, but thank you for, for your clarification long time and Claudio that let's move. So the meters can reply to questions here directly in the shot, Andreas, let's move to the third part of our session, dedicated to interoperability guidelines and to connect and to take the floor. So where we are discussing the interoperability guidelines and to provide some dates and to have three presentations so the floor is yours to manage the third part of the session. Thank you. Hello, and thank you also. Welcome to the last block of this session. So I'm happy to have with me Gabriella from Canada and young from Czech Republic. And I was very delighted to that you present part of the guidelines today here. So the first people mentioned here the agenda we start with fundamental of mere guidelines. Briefly overview about the guidelines, and then follow with the presentation from Gabriella and from young and after this, we can answer questions. We start with a fundamental of mere guidelines. The guidelines have researchers in in the middle and see the researchers and what have the researchers as output. And from this point of view, the researchers have managed Rachel published the article in thematic repositories institutional repositories, and also can publish these in journals. The researcher has mostly our research data that can also be published today in data repositories or named catch all repositories, but also in their own institution on all thematic repositories for the specific fields. And mostly the research use and develop software and this software can published in software repositories, especially in institutions or GitHub or whatever, and also catch all this for kinds of research products available. He can also publish other research things. And this could be different kind of things that we cannot have it here. And last but not least is the researchers have some information about his research project project identifier from project project is funded by funders of the project. And the researcher is related to an organization and so on and this could be fine and Chris systems current research information systems. So the landscape of a repository types are similar to what we saw before, we have an institutional and publication repositories, we have journals, we have data repository software, other products. That have it and here what coming up in the in the future and the most institutions, the moment that's our Chris systems. These in these repositories we find different types of resources in literature. We find articles, preprints, we find data sets, we have collections, clinical trials in software and so on. It's especially to say that we have for the projects and funders and also different resource types that come in the future. So, these types can found in institutional repositories in the data the Chris has mostly all of these types of these resources. And the other things data repositories can have data sets in literature, journals are mostly articles software and software repositories and so on. This is an overview about the repository types that we have and the resource types that are presented in these kinds of repositories. And the objective here for the open air guidelines is that we have these guidelines for all these kinds of resource types and repository types and is to improve the interoperability of the metadata to exchange these information, not only with open air. So, also to third parties. And it supports the fair principles to its finable accessible. Interoperability and is usable. And with these kinds. It's also repeatable by other researchers. If it have these kinds so for the guidelines using existing standards or reusing existing standards like Dublin core and data site. The guidelines are using also vocabularies and some extension is necessary. Also to reuse persistent identifier and different types and you have different persistent identify and different regions. Can we to the evolution of the guidelines. Since 10 years ago, this the first guideline version was published at 2010. Let's focus the little repositories that are mostly released or published in these in these years of 2010s. And we see that are come more different repositories up. What I say it's a special repository for data and literature and we've seen the next coming years that we have the representative of all these repositories. In 2018, in the end of 2018, we have the last major releases of the literature guidelines. It's called now institutional and thematic guidelines version four. And we have also a switch in the content acquisition policy of open air. In from beginning from 2010 until 2018, the content acquisition policy focus only for open access articles open access data and so on. Since the new content acquisition policy, we can harvesting or not only the open content also the close content and the metadata for the close content. This was a major shift from the open to close to get a view of the transformation here. And we are now in 2020. And we are in the next round of updating the guidelines. Especially the data archive guidelines, software, other research products, but also Chris in a perspective of more fundamental guidelines. And with the fundamental basis here and with the perspective of years in Europe, you know, science cloud. The last slides shows the different kinds of guidelines here, what you see in the open air guideline portal that we have. And that you find there the different types of the guidelines that are published. And with this, I would like to give the floor to Gabriela that's very excited to have some implementation and experience with the guidelines, especially for the data or institutional and thematic guidelines. And after Cabriela, Jan will present the especially the Chris guidelines. Can you see my slides? Yes. Thank you. Hello. Thank you very much and ask for the introduction. I'm Gabriela Mircha. I'm the digital repository librarian at McMaster University library. And I will present the open air compliant Canadian aggregator Canada research. To put the project in context, I want to quickly mention the project stakeholders. We have the Canadian Association of Research Libraries, open repository working group, which works in collaboration with the Canadian repository community and was assembled by cars advancing research committee. Yeah, we also have the Canada's three major funding agencies, Canadian Institute of Health Research National, sorry, natural sciences and engineering research Council of Canada, and social sciences and humanities research Council of Canada, which are interested in tracking their funded projects. I'm a member of the open repositories working group and what the open repository working group is hoping to do is to get Canadian COVID-19 related research into the open air infrastructure. We decided to start with the institution repositories. The Canadian institution repositories landscape, which is about 85 repositories is a mixture of different platforms. And unfortunately, the majority of them are not open air compliant. The ultimate objective is to have all Canadian institution repositories compliant with open air guidelines for literature repositories and take full advantage of the open air benefits, but the effort required to do this means that many institutions will not do it. It is a big commitment and many libraries do not recognize the benefits or do not have the necessary resources to allocate. So I proposed building a national aggregator to bring together research outputs from Canadian institution repository, so that it can then be made available through international discovery services. McMaster University build an infrastructure to support the national Canada research aggregator. Canada research not only solves the technical problems around open air compliance, but having a national aggregation will enhance access to COVID-19 research and to Canadian research activities in general. The aggregator is based on dSpace. dSpace has the capacity to harvest other repositories and versions five and six of dSpace can be made compliant with the open air guidelines for literature repositories version four. Thanks to the work funded by several Canadian research libraries and undertaken by 4Science. So we have launched a national aggregator that will transform metadata basically. Canada research is now live and once the records are harvested by Canada research, the local repository managers are strongly encouraged to curate the metadata in the central aggregator in order to add granular funding information in compliance with open air guidelines. We suggest adding funder identifier, funder name, funder grant and award URI. And the pros for this approach, having the institutional repository manager edit the records at the aggregation level is that it allows for granular funding information. The downside of that is that if the Canada research project comes to an end, then all the enhanced metadata may be lost, although we could offer an export back to the original repositories. So far we successfully harvested test records from several institutional repositories using the following platforms. For dSpace, we use the dSpace intermediary format, which gives us quite a bit of granularity. And then for ePrince, Content DM, Island Dora and two custom IRs, we use a simple dumbling core to harvest the data. And we also managed to register with open air. We are now several months into the project. And during this time we came across several issues in particular of course with metadata. We looked specifically at the DC description sponsorship and DC contributor fields, because these fields are more likely to contain funder name or funder ID grant or project name or number. So for dSpace based institutional repositories, because we use the dSpace intermediary format to harvest the metadata. And with that we get more granularity. Funding information from DC description sponsorship field can be mapped directly to open air funder name in open air. And then funder information that's coming from DC contributor field can be mapped to data site contributor in open air. Because for other platforms we can only get simple doubling core, then the DC description sponsorship becomes DC description. And funding information from DC description field will then be mapped to DC description open air. And DC contributor author becomes DC creator and all the other contributors, which can be the thesis advisor and so on. They all become DC contributor, which then we map to data site contributor with the contributor type other. Also, when we harvest on qualifying doubling core metadata, all the fields or all the information from the DC relation which can be has part has vision is based on and so on, ends up in the DC relation field. And because open air, sorry. The version of the space which supports open open air version for uses this relation to store the award title information. All that information from DC relation ends up in open open air award title, which is less than ideal because this relation can be used for can be used as a catch all for all references related to other items which could be year olds to other resources sometimes references and lots of references, they all end up in the award title field in when we crosswalk to open air. Also, we identified some issues around the dates. They come in all types of in many forms. As you can see, I have here only three examples, but there's more. And we don't do any transformation for the date field. We just pass it on to open air. We also may run into the situation where we have multiple DC dates, and only one of them is the publication date. The others can be embargo date can be accessed. The available date or creation date and so on. So we have to to sort through through that. And this is an example of a record from data research in open air. You can see here on the right hand side, the date for this is 2020, the year for this is 2020. But the article was actually published in 2017. So we have to investigate, go back to provider investigate where that that information was missed. We're currently working on providing the affiliation information to open air and I was glad to hear the, the discussions around that. And this is this brings me to the end of my slides and I want to thank you all for listening. Thank you, Gabriella very interesting so to see how you integrate the literature guidelines into your creator. So, thank you very much. Um, for discussion, I think it's, we have time after the next presentation. Jan. Can you please share the presentation would be good. Hello, I'm young. I am representing Europe Chris here. I'll come to that later. But let's start with the next slide please. So, which basically brings us to the fundamental question of what encompasses research information, what's the content of that concept. In Europe Chris we see three main categories for research information, it represents actors in research, which obviously are the researchers but also organizations of various types. The research process where the research projects play the main role but you also have funding research infrastructures and communication events in in that scope. And the, the typical suspect the research outputs research results. So the publications research data sets research software patents, but also you can register outcomes and impacts of research. The point is that all of these objects here are heavily interlinked and everything exists in a rich context. Well, the, the, the open air research graph is actually demonstrating that extremely well I think. Next slide please Andreas. So this research information is typically found in research information systems or current research information systems. That's where the abbreviation Chris stands for. The systems basically register and document research. That's right, but from different perspectives, it can be research done in an institution, in which case we are speaking of an institutional Chris, but it could also be research funded by a researcher or research done using a specific research infrastructure or research done in a territory where we also can see quite many national or regional crisis there. And of current research that means not just the research that is being currently performed, but it's also past research that is now relevant, or it can be the research that is currently planned. In the US you can find these systems under another acronym rims research information management systems. Next slide please Andreas. Yes, you're a Chris which I'm representing here is an association of research information management professionals from Europe, but also the rest of the world. The main product and standard here is serif the common European research information format, which is the comprehensive information model for the domain of scientific research. And it's intended to support the interchange of research information between and with crisis. Yes, the next slide. Yes, at your request we run trace the directory of research information systems. And you see we got over 750 systems registered and that is still a fraction of what is out there. So we'll be continuing in those in that effort and also making it available to open air and other interested parties through an API similar to the door, open door. Next slide please. Yes, and to support the interchange between crisis and open air we developed a profile a subset of serif to carry all this information. And here you can see the quite rich scope of that information that is being transferred it certainly involves the research outputs the publications products and patents. But it also involves the context entities for that which are the projects and funding, but also infrastructure information such as equipment and services. And also the researchers as persons on our organizations. So the institutions funders or research facility operators. And we can also record and transmit the information on the organizational structure, which is, you know, quite valuable. So speaking the communication happens on that he under the, and the typical protocol away I PMH to that zero with the exception that the payload is another kind of XML it's the so called serif XML. So that's the main difference in comparison to the other guidelines that you would meet. Next slide please Andreas. Yes. So here. So we got in beta explorer open air dot EU. That's the list of content providers that use these guidelines. It's 11 of them currently, and from quite a couple of countries as you can see. It also involves both institutional crisis and national aggregators. So this is currently in beta hopefully we get it to production soon. Yes, three cases that I would like to mention specifically three different Chris platforms that have support for these guidelines guidelines built in. One of them is the space Chris has been mentioned quite a couple of times already. It's an open source software Chris implementation. It's based on the space, but it is extended with the entities that hold the context of the items there. The main developer and service provider here is for science. The implementation of the guidelines is in pure, which is very successful commercial Chris offering by us severe it was developed by a small company in in Denmark, no part of Elsevier. One example I would like to bring up is Omega up here and offering by the Warsaw University of Technology so it's so it's focused a little bit on the Polish context and Polish institutions, but it's also quite a valuable contribution to the opener guidelines landscape. And there are also in house built institutional crisis and the national aggregators, namely, Narciss for Netherlands, and there for Finland. And from what we hear there will be other crisis to follow. Next slide please. Yes. So we are looking for content providers of other types as well. So specifically for content providers among funders where these opener guidelines for Chris managers can be used to to communicate structured information about funding at the level of funding programs and calls, but also on the individual project funding, all that fits in the in the format as well. And for infrastructure operators, it might be interesting to provide the other part of context, which is information about equipment. So these are the opener Chris guidelines. Thank you for your attention I give the floor back to the work back to to Andreas. So, I would close these block and we can open the question and answer. I would like to invite you. All of you to our feedback documents that we have for data archive guidelines Chris guidelines. So, and also it's not mentioned here, but I can give you heads to do you for literature guidelines and so on. So if you see we have an evolution of the guidelines beginning if the first guys are published in 2010. We will run the next round to update these guidelines in for different ways we see from the aggregator side of Canada. And talk with some issues here. So, I would like to invite you all to give us feedback on these guidelines. And in this time space of 10 years, there are different versions of guidelines out there for different platforms. And there's also different repository platforms out there. So, as Jan mentioned here, the space dimension, the prince. There's pure and a lot of more repository platform applications out there. And over the time, there are different versions. So, we start to create a spreadsheet. It's open spreadsheet for comment also that are refers to the different version of repository platforms in this version, and also to the versions of the guidelines. So, which guide which repository platform and version supports which guideline versions. That's the idea behind. We will share this. The link in a minute also in the chat. Let's go back to Klaus here and thank you again, Gabriela and Jan and give the floor back to Peter. Thank you. Okay, thank you. I will also ask you because there is a question here you can ask directly in the shot. Please. You can go to the question Q&A, and you can directly reply. So there was we already reply. I think all questions directly in them. And then in the chat. So thank you for all the, all the interesting questions and thank you for the great session. So long session. So we are coming to the end it's in the end two hours after. Thank you very much for joining the session. So, there are lots of interest for for for this services and for the different components of the service and of course to interact with the open air infrastructure via the interoperability guidelines. So many thanks for joining for participating. So thank you for the speakers. So many things for your contribution. So Jan and Gabriel you were the last one in the in the queue in this in this program agenda so many things for your for effort and to make it interested. So recordings and them and presentations will be shared I think some of them already shared here by the by the link. So any part of open air for those that are not and use our services providers feedback, the different by the different means. And, and, and also participate every first Wednesday of the month at half past two in the in the open air community calls we are just sharing the link here please put in your calendars. If you have any information subscribe the newsletter. Many things for your participation here. Tomorrow, we will have more information so we will have tomorrow we will have a session on targeting researchers and the services in the rdm services that we have available for researchers, please join at two o'clock open air for researchers and beyond. Our DM guides nodal amnesia are also be aware of the session and we look forward to see you tomorrow also in the in this in this in this fourth session of our open air week.