 Here I am. Okay so welcome back again. I hope you refilled. This is the last session of the day and I think it's one of the, it's probably the most important I think at the end because in terms, I'm speaking as the coordinator because it's very important to disseminate our work but also to collaborate with others which are being funded today by the same call. We're working towards the same mission which is the use. We had several zero seven projects again which are today represented by Deborah, Gergli, Raúl and Christian. They will tell us about their initiatives which are working on still horizontal but in some cases community specific efforts towards service provision and resource provisions for the use. So the first speaker would be Deborah Testi presenting DICE, so the Data Infrastructure Capacity for EOSC. It's an extensive name. These work engages with the UDOT and previous work on that and several work actions in the EOSCab as well and they will tell us more. Gergli, Stipols, a colleague and now friend, we know each other for a while, representing EGI, which again is one of the actions that offers and extends in several aspects, tool specific and community specific services for EGI towards the use and reliance is scale which are two quite important actions and I think also appealing from my point of view because they are focusing somehow on the communities but at the same time trying to offer general powerful services represented by Raúl and Christian. So I will now give the floor to Deborah, who's the first speaker and thanks again for being here to all of you. Hi Paolo, thanks for the invitation and for the organization of this very interesting event. I agree with you that the collaboration between our respective projects in this school is very important not only to strengthen our own projects but to strengthen the EOSC in general and what we can provide to the community. So thank you very much for this. I'll now share my screen if I can. Okay, thank you very much for confirmation. So as Paolo's already introduced, I'm Deborah Testi. I work at the Chineca Supercomputing Centre in Italy and I'm the project coordinator of the DICE project. As already said, DICE has been funded in the same call as OpenAir at Nexus, one of the InfraEOSCO 7 projects. So we start at the same time and we will have the same duration. We have 24 partners and our aim is related to the provisioning of the data management infrastructure to EOSC. So between our 24 partners, we have most of them that are service provider. This because also as for the other InfraEOSCO 7 project, more than after our budget is to provide services and resources to user and communities. So the majority of the partners are service provider but we engage also with some user communities and we have partners to support sustainability and outreach. As Paolo said before, the service provider that are coming together into DICE have a long experience working together. Some of them have been working together since the start of UDET. We have also some new partners, not UDET members that are helping us to support the service provisioning but with which we have collaborated already for since many years. So it's a very strong and very well connected group of providers. So which are the objectives of DICE? As I mentioned in general is to enable the data management infrastructure for EOSC. So the core of the activities is the provisioning of services and not only services but also the resources behind the services to enable fair workflow for communities. But we will also work to extend a little bit the data management service offered by working on long-term preservation policies to support sensitive data and to work with on the connection between the data services and the computing part. That will be also part of the work that will be related to keeping connection and updating the connection with the EOSC core as it will evolve in the next years. So that we will try to remain compliant with our services to the different EOSC rule of participation as they will be evolving in the next years. So a little bit more details on what we are offering in terms of services at least for in general as categories. This is a presentation of the resource data workflow that normally is being followed even if maybe not in all the steps but starting from the data collection to the data publication and basically what DICE services are trying to cover is all the from the storage of the raw data coming from the preprocessing up to when they are ready to be more globally searchable and published. If we look at the same diagram but put it into a pyramid representation so we can classify the DICE offering into different categories. So first of all we have services to cover the what we call project or personal workspace which is the short-term midterm storage that is used during the execution of a project so where the data changing and being updated frequently. On top of this we have a data archive which is midterm storage to be used in between projects or after the project has been completed so it's related to less active resource data. Then we have policy-based data archive which is mid-long-term archive again for non-active resource data which adds on top of data archive additional services like integrity checks, replication, so it's for very stable data archives. Then we have a data repository to share the final data with other members of the communities or to the generic public in according to fair principle and then we have some data discovery service to find the data that has been published. If we look at the different this type of services are very different and we have basic services so some of these services can be already accessed with a web interface. They do not require any specific configuration or authorization by the service providers they can just be used as is but we can also more satisfy more specific requests and we classify the service into standard requests and into more customized solution for specific communities or users. The difference between the last two is basically mostly on the time needed to fulfill the request and make the service available to the customer. If we put together the two type of categorization the one that I've shown before and this one we see that in DICE we try to cover as much as possible all the categories of services with as much as possible different options for user communities so to be as general as we can and be able to satisfy most possibly all the needs that will come up during the project. In the end we will have 14 services offered so we have more than 14 installation for virtual access and these are supported by more than 50 petabytes of storage offered by the different service provider. Some of the services are already available through the US portal that are the ones that have been already in production since a while but more installation more service provider will join and more services will be made available in the next couple of months through the US portal. Which are the expected inputs of DICE obviously first to scale up the offer provided via the US portal to provide scientific communities with state-of-the-art services in data management and to facilitate the open-size practice and for supporting the fair principle to foster synergy between upon European infrastructures support the collaboration in data provisioning and exchange across infrastructure and obviously try to incentivize the public actors to open up services and resources to researchers. In terms of collaboration as this is the session on possible collaboration so DICE plans to actively collaborate with in general all the European initiative but in particular we are looking forward to work together with the other O7 project and the O3 project when we start. We also would like to build a strong connection with user community we have already as I mentioned at the beginning few communities part of the project we but this is limited due to budget constraints but we plan to engage very actively with other communities outside of the project consortium in the next months. So this is a brief summary hope it was clear enough but if you have questions just feel free to ask or to contact me by email in the future. Thank you very much Deborah. So I would suggest to continue write your questions if you have any in the chat and to continue with the presentation so we can have a final discussion if this is fine with you. So I think next one in the line is Gervily am I right? Yes that's correct good afternoon can you see my slide? Yes. Yeah great so welcome everyone to this EGIA's overview presentation. I'm Gervily I work for the EGIA foundation and this Paolo nicely introduced me we've worked together for several years now on different infrastructure projects. I represent one of the compute project in this landscape and I work as the technical coordinator of this project since January. Let me start with a brief overview of the objectives so this project is very much focused on the computing aspect. The first objective is very specifically about this delivering the so-called cloud compute platform of the European Open Science Cloud offering compute capabilities to users. The second is linking to data management and to the so-called European data commons structure and vision and particularly the project aims to build so-called data spaces which is basically a collection of scientific data sets that are hosted on or near compute resources together with some analytic systems, portals, gateways, workflow systems that enables exploitation of the data by scientific audiences. The objective three is about integrating the platform with the EOS core. Actually I have some animation here. Objective three is about linking this compute platform and the data commons with the EOS through the EOS core and objective four and five are more outside looking. Number four is about working with initiatives that are similar to EOS and similar to the compute platform outside of Europe and number five is working with different providers that are not just in the consortium but wish to join this compute platform as we go ahead in the next 30 months. Similarly to the to Nexus and to DICE this project is very much focused on delivery of ready-to-use services. More than half of the budget is on service provisioning. The total budget of the project is around 12 million out of which 80 million comes from the commission and four million is contributed by the members of the EGI Federation. The whole consortium is coordinated by the EGI Foundation which is coordinator of the Penn European EGI Federation and the consortium largely or primarily is built from members of this EGI Federation and is expanding with different scientific communities who contribute to this data spaces work and act as early adopters of the overall setup. And what this setup is about is shown on this picture indicating a kind of architectural layout of the different layers where we position our services. We start fundamentally with the low-level infrastructure services which consist of different research clouds operated by universities, research institutes, public clouds operated by commercial entities, high throughput compute systems, clusters that are designed for massively parallel applications and high performance computing systems that are supercomputers that are suitable for closely coupled applications. On top of these infrastructure services we have different kind of federation tools that enables and allow users and user applications to use multiple infrastructure clouds, multiple clusters, multiple supercomputers together. This is achieved by providing federated computing, job distribution systems, federated identity management, a single sign-on across systems and federated data access enabling access to distributed data from any of these compute systems. As we move closer to the to the user layer or the top layer we have more user-facing capabilities such as interfaces, web portals designed for interactive computing, automated deployment of scalable clusters in these underlying systems, different services designed for artificial intelligence and machine learning applications and frameworks that help user portals distribute a high number of jobs and tasks across the available capacities. On top we have these data spaces, I will have further slide about that one. Basically these are the science discipline specific capabilities that are offered for certain disciplinary areas that incorporates applications, visualization tools that are designed and coming from the discipline and they exploit the underlying compute and federation capabilities. All these services are connected together through different service management tools, processes and policies by which we mean different monitoring systems for example that observe the availability and reliability of the services, accounting systems through which we can observe who used what, how much consumption was made by each community or use case and the documentation that the different service providers use in order to ensure responsiveness towards the use cases. All this tech is delivered through the ESC portal and made available for ESC users. And let me just show you one example that combination of these elements is possible for a given use case or for a given setup and for example the project provides so-called notebook service which is some of you may know is based on the Jupyter notebooks which is a web-based interactive computing environment through which one can write scripts, one can write codes and mix that with documentation, mix that with multimedia content. The CGN notebook service which is supported by the project offers access to big data that's distributed across the network, can be sitting in different scientific repositories and can import those big data into this interactive computing environment to perform to define and run analysis. Another service that this whole use case can rely on is the GitHub and the Zenodo repository that many of you know from the previous presentations and from the previous years. The setup enables users to export their notebooks and store them for the long term and share them through Zenodo and then other users can reproduce basically the same analysis by importing those notebooks from Zenodo into the so-called binder service that EJA offers and then import the same big data and rerun the same analytics again or in a customized way. So that's an example of how the cloud compute services, the different federation services and the interactive computing capabilities are all integrated together and work together also with GitHub and Zenodo to reproduce big data analytics. The full portfolio of our services is quite large. We have 61 providers and they contribute to 31 different types of services. You can see here that the largest contributor list is under the cloud compute service indicating that major part of this project is delivering infrastructure as a service cloud compute capabilities. This cloud compute enables the users to run their own applications to set up their own system within remote servers. The other services that we provide add extra capabilities on top of this underlying infrastructure. Many of these services of these 31 different types of services are already available in the Yask portal. If those of you who know the portal in the marketplace section there is a so-called communities and infrastructures grouping and within that you can already find the EJA as a community or grouping and if you click there you can see the 25 or so maybe a bit more services that are already available out of these 31 that I showed earlier and through this you can either access the services immediately because there is already ready to use configurations or you can request your own setup from the provider. One of the recurring questions is how capacity is allocated and how it's managed by the project. We have a large number of capacity providers who provide this cloud high throughput compute and high performance supercomputers and they assign this capacity on one hand to communities that are already in the project and I will talk about them in a minute and obviously for new users because that's the point of the whole project to be open for new business. The total capacity you can see here is distributed between traditional CPU based computers GPUs and also storage that's available for mid or short term usage. It's basically a scratch base that one can use while doing a certain calculations. All these compute capacity is allocated through a mix of funding mechanisms. On one hand we use the virtual access funding model and we have 16 providers using that one and they are paid by the commission to deliver the capacity so the access is free for any new user community. These 16 providers have been selected during the proposal preparation based on their location and the type of compute flavor they can offer in order to be able to cater for diverse needs but we also have agreement as a project with additional providers who are not paid by the commission but who are paid by their national government or funding agency or ministry to deliver compute capacity through EOSC for different types of audiences and integrating those policy based access providers enables us to extend the total capacity that we can assign for new use cases. We need to consider those different policy constraints that they have. Very often they are open for users who are coming from their own country but doing this intelligent brokering and the allocation of capacity we can hopefully satisfy all the incoming use cases. We also have pay for use providers, cloud for empty systems, offer capacity for payments so the payment on this case is for the users and we have also four sites that deliver supercomputers for piloting activities in the project and how these capacities assigned to users is shown on the on the next slide. We basically work with or expect to work with two types of use cases. One is that kind of long tail of science use cases the smaller user groups or users who want to try out things before going into more serious business. They can come in and they can access the capacity through the EOS portal. We have ready to use allocations that is open for many of the users and on the other hand we expect to have more complex use cases from multinational international communities where we need to fine-tune the setup. We need to allocate significant capacity in given set of countries or sites. We aim to attract those use cases and serve those use cases through a call that we will publish very soon in the coming weeks and we plan to have a cut of dates or evaluation three times per year for the submissions and through those evaluation process we will assign providers to each use case who will serve the use case and support them for aiming for the longer term and customized engagement. The scientific communities that are in the project are represented here. On one hand we have these data spaces that I mentioned is basically a mix of infrastructure capacity services storage and compute together with domain specific libraries and applications all combined together and offered as an integrated setup for example typically as a web portal through the EOS portal. We have 13 of these data spaces you can see here the scientific distribution and also the names of the different communities who offer them very often they are linked to as free research infrastructures or similar international initiatives. We will not just deliver these data spaces services but also perform a fair maturity assessment on those in order to ensure that they not just support users but support scientists in such a way that the final outcome of the research those researchers conduct are meeting the fairness criteria of EOSC and enrich EOSC with new scientific outcomes. We also have other types of user communities in the project we call them early adopters they basically are less mature they are not aiming to become providers in EOSC they want to use the EGIS services in order to carry out certain computing activities. And let me close with some notes on possible interaction with Nexus and with other 07 projects we had the first similar meeting at the EGIS public launch event so my ideas or our ideas are based on the knowledge there also I slightly fine-tuned these messages based on what I heard this morning. So one is I think we should work together on closing this whole research cycle and make this research cycle as easy to carry out for researchers as possible. Paulo started with this virtual cycle in his presentation fortunately I couldn't take a screenshot I wanted to include something like that basically here the idea is that EGIS would fit into the I think the left side of that slide that Paulo presented more in the computing and short and mid-term storage space and we should work with DICE with Zenodo with Nexus on importing scientific items into this computing model and pushing these final scientific outcomes or computing outcomes in the form of publications or in form of datasets into these longer term archives and repositories and catalogs. The second is working together on tracking the impact let's say impact of EGIS for example from my perspective or ours perspective this could be based on the EGIS commit dashboard which we have in a beta version I think we should move forward and move that into the latest dashboard and tools that you presented this morning and we probably should discuss how to do that and how to how to customize it in such a way that we can capture EGIS impact and we can capture EGIS impact and we can capture EGIS users impact in that you mentioned earlier that some of these will be part of the EOS score or are part of the EOS score I think we need to understand what are those and probably start with those as the EOS score is expected to be the long-term entity that we need to interface with. Very specifically with the Earth Observation Project C-Skill and reliance we work or we have worked on hopefully on the integration of their services and data with the compute platform that EGIS provides. Number four is to align the support for the big communities there are or there is common communities in some of the projects that we work in I think we should seek for these these commonalities as we go ahead and sign up and support new use cases. One of the activities that I mentioned EGIS will conduct is this so-called fairness assessment of the EGIS data spaces maybe some of you who did something similar could help us in here. Number five is the grid computing which is about lowering the environmental impact and CO2 emission of the overall compute platform that we deliver. I expect that your project has a similar activity there I think that's something worse to be aligned on what are the practices that you will apply and we will apply in order to have some joint activities here and the number six is I think the simplest is cross-project promotion and dissemination of the results of the opportunities that we all make available for ESC users. So as a summary EGIS delivers the compute platform which consists of basically two layers one is the compute continuum with HTC cloud HPC machines and related platforms and there are these online data analytics services the data spaces. We are already open for business most of the services we deliver are accessible through the ESC portal and we will have call for use cases published very soon with the first so-called cut-off date or selection to be later in the spring. Next week we will have a meeting with with Paolo and some of your colleagues from open and access so I think that's will be the the real work where we start the joint endeavor. Thank you. Thank you very much. The last thing was precious and I think aligned with what many of us have in mind but we concrete ideas so thank you very much. So shall we continue with Raul? Yes hi can you hear me? Yes we can. Hello Raul thank you. Thank you everybody and especially Paolo for for putting this together. I am very happy to be here as well. Okay so for everybody I mean my name is Raul Palma. I am working at the Postnat Supercomputing and Networking Center in Poland. This is the largest HPC provider in Poland and I am the coordinator of Reliance Project. This is one of the zero seven projects as it has been mentioned before but we are in the topic of A6 which was about extending the research enabling services in EOSC. So at a glance in short Reliance I mean first of all it stands for this long title but basically it captures what we are doing in Reliance so it's the research lifecycle management technologies for earth science communities and Copernicus users in EOSC. This is what we are doing in a very short title. Now in particularly Reliance seeks to extend the EOSC capabilities with an enhanced support to different research activities through a set of different interconnected services which are quite innovative and in alignment with the EOSC Interoperability framework which actually is a few of the authors are part of our team. Reliance also through the adoption of a holistic research management approach it aims to enhance the discovery of and the access to research data and resource with particular focus on Copernicus data improve the extraction of relevant information managing the research lifecycle while promoting the first and open science principles. So this is of course very much aligned with other projects of zero seven but we are going into a higher level as I will show you in the next slides. So our pilot I mean our project will pilot and demonstrate the services and their value in three earth science communities and we will engage others via some open call with the goal especially to foster the use of Copernicus data and to enhance in general the support in EOSC for multidisciplinary research improving the EOS science as a whole. The priority is it's a shorter compared to the previous one so we are 24 months we have a smaller budget is to two million euro budget but we started together with the others so in in January 2021. So the service portfolio of Reliance is based on three key and complementary technologies and approaches that these are basically what comprise the ecosystem that we are planning to deliver. So first of all we have research objects which is used as the overarching mechanisms to manage scientific research activities and connect associated resources. I will give you a bit more of context of research objects later on but the technology behind this is called ROHA so this has been a project we have been working for several years already and we have over around 3,000 research research objects in the in the repository at this moment. Next we have some data cubes which are basically providing the mechanisms for enabling efficient and scalable access to and discovery of data and access particularly air observation data so in the platform behind this is called Adam which already includes several petabytes of data coming from Copernicus. Then we have services for text mining and enrichment that will allow to extract machine readable metadata from resources in the research object and other researchers for example to discover scientific information at scale structure their own research. The technology behind this or the service behind this is called Kogito from one of our partners in the in the consortium. So based on that we can say summarize the objectives of reliance in the following. First of all when we want to provide a suite of services for the research life cycle management there are based on research objects data cubes and text mining which are in compliance with fair principles. We are aiming to integrate on board of course all of our services into AOS leveraging and connecting with other complementary services that are already available there. We will demonstrate and validate the value of our services through several use cases in real life scenarios. In particularly we have one multi-disciplinary use case and five thematic earth observation use cases that I will also briefly mention at the end of this presentation. And finally our goal is to foster the adoption of Copernicus data and AOS services in such communities to increase the impact of this in the whole European context. So in terms of the project partners the consortium is smaller so we are just nine partners from five European countries. Particularly we have three research communities involved there so these communities are from the earth science I will show you a little bit later about them. We have five technical partners and four of the partners are from the industry and you see in the in the slide the geographic distribution. Okay so being as one of the background technologies in reliance it's important to give a bit of overview of research objects and probably some of you already know about this and we have heard even the word research objects through the presentations in the past. So we started to work with this for many years ago from 2009 more or less in a project called work for forever but this has evolved of course during the years and several branches have started to spread in different directions but all in all we are still having a community about research objects we are having probably also you some of you know this arrow crate community that we are part of of course as one of the initiators of this initiative but research of this is basically a rich information object that accounts and describes and shares everything about your research including how things are related to each other in a way that is understable by both humans and machines. They are of course you can think of them as a kind of logical container that has a unique identifier and that connects different kind of resources that can be for example hypotheses the data that was used the results that were produced the method that you were employing and adopting scientific workflows the provenance versioning information the people that was involved and annotations about all of these resources. Research objects have been used and demonstrated in different communities in previous projects from bioinformatics to astronomy to earth science and as a result they are becoming like gaining more and more attention as one of the promising research enabling technologies. So based on that and having this into consideration in this picture in this diagram we can see a high level view of the reliant service architecture and their connection with AOSC and other existing services. So first of all you can see in the middle we have the services of reliance that are interconnecting and complementing each other and that will enable researchers and scientists to to access all this functionality provided in particularly their work from different interfaces using the research object as the main connecting point. So all of these services will expose some restful API some we will also generate some Python libraries that will enable also the communication with different services in AOSC for example but also the communication and access through some client applications and in particular it's important to know that data cubes are going to be linked in the research objects as first class entities so they will be described with a rich set of metadata enabling the you know not only the access to this data that they are providing especially like Copernicus and so on but also to enable the reproducibility and reusability of these access mechanisms in the future. Then we have all of these services for text mining and enrichment that will be automatically incrementing and enriching the the research objects metadata for example extracting information from resources like the data cubes annotations but also about the resources scholarly communication resources in the research object and and so on. Some of these connections are already in place so you know we are not of course starting from scratch but we of course need to make some adaptations to connect to AOSC as well. And also as you can see in the diagram we are leveraging of course some of the AOSC services from the cross core cutting services but also some advanced services that can be of added value for our researchers. So these also include other existing services like this scholarly communication services you know Jupyter notebooks from like like the EGI notebooks that was mentioned before but also QCG services for HPC management resources and so on. So based on that in the overall picture you can think of reliance as playing a complementary role to what is already in place that is already available. So reliant services the idea is that they will act at a higher level as you know as a bridge between different AOSC services and resources that are already available. So the idea is that they will be able to connect for example from the data that was used by researchers for example the data cubes but also other data in AOSC services or repositories methods that were used to process the data that can be accessed via you know for example the EGI Jupyter notebooks also the research infrastructure that were used to to you know execute all of these calculations or research in general to the results that were published in scholarly communication services and repositories and other open air and so on. But overall as well taking into account the research lifecycle as a whole entity. So for example during the research lifecycle we will be able to create snapshot releases that will increase and make faster the the publication process and for example we were connecting to again repositories like like Zenodo to to deposit whatever is being produced and released and at the end of the cycle for example when the final release of the research is made this will kind of bridge and close the gap from the moment of the data that was gathered and access to the moment of the publication has been released. So very briefly I want to introduce our research communities so we are that we've been validating our services we're bringing three main earth science communities that are working in different aspects and different disciplines. First of all we have the sea monitoring community that is studying the marine habitats and the present past dynamics and the ocean and the interface between ocean and atmosphere. Then we have the this community first of all is represented by CNR East Marine Italy they have a very large community of user community of researchers and they have a lot of connections with others. Then we have the Geohastar community that is aiming to provide and improve the scientific knowledge of volcanic and seismic supersize to support the disaster risk reduction. This community is represented by INGVU also in Italy but they are also being represented as one of the key partners of this Geohastar Super Science and National Laboratory where they are sharing the scientific advisory committee so they are very strong in this community. And finally but not least of course we have the anospheric and climate modeling community that includes scientists working on the development of numerical meteorological ocean and climate models. This community is represented by the University of Oslo the Department of Geosciences. So these communities will be driven the one multidisciplinary use case that will be working on the the impacts on coastal environmental during the 2020 lockdown. So this will be a very interesting thing that will encompass many different disciplines and they will also have five thematic use cases. So overall this will showcase the reliant services and how they can support researchers and promote the cooperation among scientists in earth science. But it's also important to take into account that these services may be adapted to any other research domain. And with that I just finish and thank you very much and of course I am very happy to have you on this. And thank you Raul. I think we better postpone the questions at the end. We also have two parallel sessions for those who are willing to participate. I think it would be really worth continuing on several aspects the discussion if you have time. So we are now getting to the last presentation from Christian. So leave it on the floor, please. Yes, thank you. I hope that you can already hear me. Yes, we can. Here, see you and presentation. Also these slides are theirs and then I follow on. So as a last presenter now I would also like to present our activities within seascales of the Copernicus EOSC analytics engine. We are also focusing on earth observation and have the big ambition to build an easy to use platform to make use of the huge amount of Copernicus data that is available. I'm the CEO of the earth observation data center situated in Austria and I'm also in the role of being the project director of seascale. Yeah, within the consortium we have a number of partners. I will present them afterwards in one of my subsequent slides. What is our problem statement? So the European Copernicus program has really established itself as a predominantly predominant spatial data provider in providing more than 10 of terabytes of data per day that basically gives insight in the current status of our earth. Currently, we don't have any single European processing backend that serves all the data sets of interest towards the scientific community that really hampers at the moment the use of the data and makes it really complex. Either you use some of the U.S. services where you maybe have also quite a number of limitations and therefore our and also maybe concerns to use them. And therefore our big ambition within seascale is to provide a big data analytics federated infrastructure with with cloud computing and storage architecture in order to really quickly use and analyze the long term archive that needs to be established for the Copernicus data that we all require. So therefore we will set up seascale in a manner that we integrate it with the help of several researchers and institutions over Europe to build up this federated infrastructure to provide this infrastructure within years so that scientists can make use out of it and enable state-of-the-art research towards subsequent use of observation data for other scientific disciplines for strategic European decisions, facilitating the European Green Deal ambitions, facilitating also the new initiative from the Commission called Destination Earth where basically a digital twin of the Earth shall be established in order to analyze processes on the Earth's possible future influences and looking at and compare the current status, this past status of the Earth. So that's where we want to contribute. So the high level of objective of seascale is to integrate this federated computing and data platform into EOSC. We will include the users from the very beginning in order to set up a system that is really perfectly fits to their needs. So we will have an architecture blueprint, we will have a constant onboarding process to know new users but also to new service providers so that this federation really also goes beyond the project partner. So that's really one of our goals. So we will not limit ourselves within the project. We want to set up something where others can contribute from an IT perspective, from a data perspective, but also from a user perspective. Yeah, so where did we start? We had a look on the current situation in Europe. So we have the diocese that have been established by diocese projects that have been established by the European Commission. We have national initiative so-called color drift ground segment nodes and other activities. And on the other hand we have the infrastructure services that are well presented within EOSC. And our idea is to federate these two groups towards a large scale infrastructure and really bringing together the expertise on one hand from the EOSC sector, with UDC, Dildares, Vito, Calcfaro, Tiovin, Zesnet and Gionet being involved in the project, but also with infrastructure providers. Some of them are identical but we haven't reached the consortium by EGI as an important player by NFN, SurfSara, and also INCD. Gionet and Zesnet are basically acting on both sides already. And the aim is really to tighten the EOSC sector with the EOSC and the infrastructure sector in a very close manner. So that's the big C-scale ambition. How do we or what will we provide? We will provide this in EOSC data archive consisting of current but also historic long-term data sets from Copernicus and other satellite data. We will provide compute services which means on one hand cloud environments but also HPC services that can be used for the analysis of the EOSC data. And we will provide an analytical platform on top of these resources so that an easy deployment of solutions and use cases can be performed on top of this federated activity. Yeah, all these services will be integrated into the EOSC portal with the ambition to do this somewhere mid next year so that everybody can book these services and can make use of these services in these scientific questions. Yeah, everything within C-scale is significantly built around the use cases. So we have defined six use cases or use cases already within the project so that we have the users basically on board and they drive the development but we will also onboard other use cases within open coal where we also provide virtual access and free infrastructure costs for them so that even the use of the platform will be significantly widened in respect to what's also new ideas and new research communities. Yeah, the idea is really to validate and optimize C-scale during its development phase towards supporting an agile approach, getting user feedback immediately considering their feedback in their current development. So that's one of the ambitions. Yeah, the solution should be really cloud agnostic, independence of commercial closed and non-new providers so that's what we want to achieve, cross and interdisciplinary exposure so we don't want to see Earth observation and the use of the Copernicus data just in a limited field. We really want to make it as accessible and this access should also be done in a fair manner. Yeah, in the following I just want to present you three of the use cases so that you get an idea of what shall be possible. The first one is the aqua monitor that is currently running on the Google Earth's engine in a quite successful manner. We want to transfer this towards EOSC and the infrastructure and data that is provided between C-scale. So that's one of the big ambitions that are around to move that from the Google Earth's engine towards EOSC and having that around. Yeah, the second one is water watch, a near real-time update of the surface water resources, so lakes, rivers and wetlands. Also here we want to have the real-time monitoring available within C-scale. The third one is then high sea, again about water that's maybe also related to the significant involvement of deltaras, but instead of the monitoring of the water areas it's here now the big topic of water quality at the seas. Again here we will require a huge volume of data. We need to have a lot of computational capabilities in order to discover monitoring of that high sea activity. Use case number four, five and six I don't present in detail in order to keep the presentation short. So high resolution land surface analysis that's within the high-res LSTA use case. The big topic, the second use case number five is the return use case from the University of Wageningen is focusing here on the monitoring of tropical forest areas with the help of multispectral satellite data as well as radar satellite data and last but not least the virtual European Sentinel data cube where we want to provide a new access possibility to higher level Sentinel data that then can be directly used for the implementation of other use cases. So that's the six use cases that are already fixed. As mentioned we will also provide a number of virtual access resources from different providers so that's the overall numbers we will have more than 1500 CPU core hours a year. We have significant HPC resources and we have more than a petabyte of storage that we will offer towards users that they can book and that will be provided by this virtual access mechanism. Yeah to sum up already so the big ambition is to federate the infrastructures with EO data, data services that are around and infrastructures that are around and to make really Copernicus resources easily accessible towards research areas and the EOSC in general. We will deliver this federated infrastructure services so the data, the compute but also the analytics service that we have presented that shall all be accessible to the via the EOSC portal. One important factor for us is that the federation will be co-designed with the researchers so they will be involved from the very beginning in the setup and we will make this environment then available to others via the virtual access mechanism that is provided in the call option and yeah we are really looking forward in order to implement this C-scale idea. Yeah then I already thank you for your attention here you find the contact information so this info at cscale.u you can get all the details you can also register and ask for the questions there we are currently setting up the webpage and that's still under construction but also will be available soon to the basically whole world so that we can inform people about our C-scale activities there was a Twitter hashtags around the C-scale in order to receive updates via social media. Yeah with this I close now really thank you for your attention and I'm happy to receive some questions however I need to leave in five minutes latest. Oh sorry I was muted thank you very much. So I think there is a lot of material and very interesting material as I anticipated so I tried to take some notes trying to share them now with you okay can you see the slide? Yes yes okay so in some cases this has very personal notes so you won't find what I need to buy for my refrigerator but some of the things that I surely would like to discuss with you so do you first of all is there any questions for the speakers? Things that you'd like to be clarified or that are not clear enough okay so from your presentations I see several possible collaborations bilateral in some cases which can be extended to a more general purpose approaches so this is the second bullet that came up to my mind so we can of course consume each other's resources right especially when this is a specific action like data storage or computing provision then one project where this is not already the case for example EGI is part of reliance one project may benefit of the resources of another so you can apply for the tenders you could do the actions that are necessary for that but the second point is make our services interoperable okay and I see many opportunities here one is bilateral which I think is interesting of course because we could benefit from each other's services but it's the one that I less prefer in the vision of the years so I would like for example any interaction between two services across our project to be driven by the broader mission of defining protocols that would allow other similar services to perform the same interaction so on this respect for example and this is a question for DICE and the EGI years I see an opportunity there for you to exchange for example the results of your processing in EGI for all the use cases that you have to be fully integrated with data publishing and also with what we're providing in open air so the overall publishing in the broader context so what do you think what's your opinion in that direction wants to start Deborah maybe yeah yeah I was thinking more or less along the same line this morning when I was hearing all the different presentation on the open air services so for sure one one possibility is to rely on the open air services for the publication what comes from the DICE services or from the computing I'm not to talk about EGI but he can speak for himself but as a general picture I think that would be very interesting and would provide a complete set of services and from the data processing data creation data processing storage and make them available to the public across the three project at least maybe involving also the first thematic issues but for general purpose I think in general purpose services I think the three project can find a nice solution for for user communities yeah yeah I agree yeah that's just a kind of analysis where I think we should yeah I mean yeah then we we should make a technical analysis I agree but at least in in in principle I see it a nice has a nice idea so would you agree on the fact that we should aim at more general purpose frameworks so going towards rather than bilateral integrations from broader yeah then maybe the general purpose framework will lead to some bilateral work and bilateral integrations but use cases yes would be bilateral with the expulsion of applications of this but I think this is the best we can do for the years in general right so the more we do it the better so we we have I like to hear we're all aligned on this kind of perspectives I'm asking also to the colleagues from C-scale and reliance if they want to mention you want to have a voice on this yeah hi just easy role well regarding then from my side of course I see a lot of potential collaboration and you know reuse of different things that other of the projects of 07 are doing for for the one hand as I tried to show on the presentation I think in our case we want to you know leverage existing data sources data repositories like the ones from v2share or some others that may come from from dice but also repositories like synodom that is coming from open air nexus and on the this is also the computing infrastructure like you mentioned the dji jupyter notebooks for example this is very important for us as we are planning to use this kind of environment as the default computing environment for the research object let's say so we want to create some kind of libraries pattern libraries that will be able to access the resources in the research object and do whatever you do in your in your library but in your notebook but then you can always come back to your research object and then after that of course we have the part of the publication site like I mentioned this is very important as well so our research objects I mean we have our platform but what idea is that we can actually publish this in some of these existing scholarly communication platforms and this includes of course again synodom because it's very popular one so for example the idea now that we have is whenever we are creating some kind of releases of the research object that we have like life cycles then we we will have the possibility to release this also into synodom so we will just be able to deposit whatever the user was the researcher was working on the the team of researchers were working on the research object as also one entry in them in the synodom and this will be able then to we need to further discuss this and we are in strict contact with cattle gobble on these kind of topics on how to nicely and to the point integrate research objects in synodom and in here because I think it's a key aspect of your work as well so it's good that this closes the circle in a way Christian from this scale uh paolo christian wrote in the chat that he had to leave okay yes he said five minutes sorry I should have asked him first okay so we now have two parallel sessions I hope many of you will be able to stay to further discuss these aspects because it's really important for us to receive feedback can I share my screen paolo so that the participants can see the the different sessions no as you can see they have two different connotations one is more on uh nexus services in the national setting so how can we make sure that our services are useful are useful at the country level to support the uh dissemination adoption of take improvement and channels regarding open science in general open science practices and this has a regional say national perspective because uh open air has representative at the national levels which are trying to cope with the different maturity level and different expectations in terms of technologies road maps which the national level are different so what we would like to discuss is to how to serve these people uh which are helping the open the open science dissemination in the best way so this is the first session let's say the other parallel session instead is more related with research communities and the rise so how the services can be better customized to serve the needs of research communities and the rise and that's uh of course in collaboration with other infrastructures in collaboration with the research communities etc this is this has more a more technical again um bias but still uh policies are in place uh to the point that we can uh uh flexibly adapt to them and understand what they are at a community level it's a different level of customization we'll say uh so we have two rooms uh how can we access these rooms uh Paola the first one for no words I think they already have received uh an email well for for registration so as it's a closed session um it's um we shared via email the link for registration so they maybe have just already received a reminder as for this session on research infrastructures you will stay here in the same in the same room okay so sorry for the the missing this uh the time I I get it wrong because it's 50 okay it's it's happening now we we won't have any any breaks so for the ones for the no words we want to to move to the other session please check your email um for the reminder of the maybe you can maybe you can paste here the link to the other the link for registration yes yes I can pass or no or the link oh yes yes you need the registration otherwise yes yes for zoom yes