 Labs working on onboarding data at scale onto the Filecoin network. He is focused on getting public data sets onboarded to the network via the Slingshot programs and is also part of the Filecoin Plus governance team working on client onboarding and trust. Reach out to him at any time to chat about data cap, DAOs and some excellent puns. Today, Deepol lead us through Filecoin Plus and data DAOs. Welcome. Thank you. I hope everybody's doing well. Sorry I couldn't make it there in person, but it looks like a fantastic event and really well set up. So yeah, please have a blast today and celebrate a little bit extra for kicking ass and getting together and doing productive things in my stand as well. Thanks for the excellent intro. I'll probably just skip this slide in the interest of time. Just covering, you know, leaving this up mostly for the bottom where if you need to get a hold of me, find me on Filecoin Slack as Deep Kapoor or find me on Twitter, Deep K Kapoor. Happy to chat about any of the things that I'm going to be discussing in this presentation or anything at large with regards to large scale onboarding of data under the Filecoin Network. So what I want to cover today with you in the short 20 minutes that we have is number one, Filecoin Plus. For those of you that aren't familiar with it, we'll go over the overall sort of core concepts of the program, how it works, how the pieces sort of fit together. We'll talk through Slingshot and the various Slingshot related programs that exist today. And at the end touch a little bit about how these things sort of are moving towards the direction of doubt or doubtification. I would say sort of a larger scale understanding of what it actually means to do that and then how you can get involved and what's next. So without any further delay, let's talk about Filecoin Plus. So Filecoin Plus by definition aims to add a layer of social trust to the Filecoin Network to incentivize useful storage. What does that actually mean? Great question. So we're going to say a bunch of things here and I don't expect you to understand all of them, but then we're going to dig into each of the components and then hopefully it'll start to make sense. So the first thing I want to tell you is that in the world of Filecoin Plus, we have these different stakeholders. You can see them in colored boxes over here. I don't know if this screen is big enough for you to see this in person. And so I'm just going to walk you through it because the text might be a little small. So you have community governance, root key holders, notaries, clients, storage providers, and miners. So the idea is that root key holders, notaries, clients, storage providers interact through the allocation spending of a resource called data cap. Effectively, notaries serve as a key stakeholder position and we'll dig into each of these by allocating data cap to clients. Clients spend that data cap in storing deals on the network. So specifically clients who are data owners who are bringing data to the Filecoin network use it to make deals. And this will start to make very soon. But then in making those deals, storage providers who receive deals that are sort of coming in with data cap receive a boost to the adjusted power on the network. And so you're at a crypto account conference. It's important for us to explain to you what that actually means. The way Filecoin distributes its rewards is based on power. And if you're taking deals that have data cap on them, then you're getting a 10x boost to the relative power, the relative footprint that you're occupying on the network with regards to that deal that you're taking on. So if you're taking a 32 gigabyte deal or a gigabyte deal in this case, it looks like 320 and inflates the rewards that the network is giving you. So in essence, this is a network-based subsidy of dealmaking between data owners and storage providers. Now let's dig into what each of these different stakeholders are, define data cap, and sort of help this all make sense and why it's interesting. First things first, what is actually data cap? It's a novel, generated resource within the Filecoin network. It's programmatically generated. We'll talk about how that actually happens. But as I just mentioned, clients and data owners use data cap to make storage deals. It's a one-time use credit. So it's consumed as it is used in those deals. And then the storage providers receive a 10x quality adjustment to the power for serving deals that have been made with data cap. So back to the sort of core stakeholders of Filecoin Plus and sort of defining them a little bit so you understand how they all come together. First, we talk about the root key holders. Root key holders are executors for decisions made by the community on the blockchain network. So they're reputations, sorry, they're representatives of reputable organizations that are in like our ecosystem effectively, Filecoin, IPFS, Web3 sort of spaces, but ultimately what they actually are are their signers on a multi-seg. And that multi-seg has the power to invent data cap and issue it to notaries. And notaries being other addresses on the network which are then flagged with the identity of a verifier. So just recap, community figures out who the notaries are and we'll talk about that in a second. Root key holders are the ones responsible for executing those decisions on behalf of the community. This typically means assigning the value of data cap to a notary that they can allocate as well as the status of being a notary on the network. Notaries are arguably the most important stakeholder that keeps this engine going. Their role is to do KYC and due diligence on incoming data owners and clients out of the network. The goal is for them to identify trustworthy clients. So this goes back to the top line sort of definition of Phil Plus, which is that it's a social trust system, right? And so notaries are the ones that are actually looking to establish that trust to figure out who the clients are that are worthy of receiving this data cap that will give them these extremely free or cheap storage deals. And so the idea here is notaries spend whatever effort they need to ensure that they build a relationship with the client, understand who they are and give them data cap and recommend trade with a level of trust that is warranted based on the information that is provided by the client about what they're working on, who's funding it, why they're doing it, what their goals are, what the long-term plan is, et cetera. They also have a set of operational guidelines which includes how they should do these things, toolings that needs to be used, need to keep a record, bookkeeping for auditing purposes and ensure that they can actually talk through past decisions that they may not be held accountable by the community. They're selected through an election process which includes putting an application out in the public which is scored on a rubric. So that rubric is also open source and the idea there is the community can iterate on the rubric, notaries are scored against the rubric, we do a pretty open election process where a notary applies, it's scored on a rubric, gets to sort of contest and then the community also gets to contest and then we finalize the rubric scores, rank all the notaries by their rubric scores and then pick a set as defined by what the goals of that election process are. So hopefully this is starting to make sense. We elect notaries as a community, rookie holders give that particular address on chain, the ability to do these special things like issue data cap to clients. And so clients, these are the actual data owners as the demand side of the storage marketplace, they're aiming to be and come across this trustworthy so that they can get this resource to make extremely cheap deals. Storage providers represent the supply side, as you can tell by now that loss and certainly not least is the actual governance community. This includes of course all the above stakeholders but also includes just anybody interested in the community. Our governance processes are completely open. We use GitHub as our primary tool for making progress in the program, indexing, documenting and then we do open governance calls bi-weekly. And so literally anybody could participate, includes you and come to our calls and help with things like defining the rubric, defining the roles, changing, adjusting the roles, putting in FIPS to adjust the way the program works, et cetera, et cetera. Quick recap on the flow again. So, notaries perform due diligence or KYC on clients to verify them. So you'll often hear the term verify but really it's just established trust. The real client, they're looking for real clients with valid data that would actually bolster the usefulness of productivity without the network further as goals in like being the marketplace and search interaction point for all of humanity's information and ensure that the subsidy that's being created crypto economically on the network is actually being utilized in the best way possible for long-term, whether that's for the actual clients and their ability to do things because the non-profit organizations, scientific organizations, et cetera, or if it's for the network in terms of, hey, this is a really compelling use case that we want to serve and we're excited in making sure that we're able to deliver this for everybody and show them that like this can be done on Filecoin. It may also include like stretching the network to understand how to handle. So we've got a working group, for example, that just spun up, they specifically think about enterprise support in Filecoin Plus. So how do we deal with data sets that can't be open all the way or are encrypted or have some private components? Like how do we still build trust with clients? Because traditionally sometimes, notaries would just ask clients for samples of the data, but that's not necessarily possible in this work. And then based on the identification of clients and their reception of or allocation of data cap, they then go expend that in deals. As they burn that data cap, storage providers get adjusted power increases and get better rewards from the Filecoin chain, as long as that deal is live and that sector is running, et cetera, et cetera. Clients have three parts to getting data cap today. They can go into automatic verification, good example, this is a site verify.lift.io. For anybody with a good hub account, greater than a year, I'm sorry, the greater than six months old, I think it's one of your days, yeah. You can pretty much get 32 GB every month. We're adding a scaling component for this as well, based on how you're using it and how frequently you're asking for it and show that you continue getting more if needed. This is extremely useful for testing purposes. Like if you're just wanting to get your first deal or get your first 20, 30 deals, see how it actually works, build up a system, do a mini POC for yourself. This is the recommended path. If you're a client who has a single use case and not a massive dataset and is just looking forward to getting this done with the least amount of effort and the least amount of operational overhead, you opt for the second path where you typically get between 10 and 50 TB bytes of data cap. That's through working directly with an electric notary. All of this happens in public, all of this happens in GitHub issues, but this does have the option of, oh, can I email you this stuff instead? And then you maintain that in your books instead. And so it has a little bit more flexibility for clients. And then the third, and by far currently the most popular option is to go via this, what we call the large dataset and repath or like the Filecoin Plus for a large datasets program. And here the goal is you're requesting 500 TB plus for data cap. Each application can be up to five TB, but it's released in tranches as opposed to that data cap going out on its own. Completely public process, much larger, accessible amount of data cap, but because it's released in tranches it's a little bit more scrutiny. Every tranche that's released requires some participation from notaries, et cetera. So where are we at with the program today? So on the right, you can see this graph. This is the last three months of active deals and heavy bytes on the network. The light green portion on top is Filecoin Plus, the blueish portion on the bottom is like regular deals. You can see that Filecoin Plus has really been on a tear since the new year. The LDN process has resulted in a lot of clients getting data cap and onboarding tons of data at a much faster rate than the way before. We're currently at the point where about 75% of all the data stored on chain is actually verified data coming into Filecoin Plus. That means more than 55 PB bytes of data cap being used in deals. And that represents about 75% of the data cap that's been given to clients. And so there's still about 25% that's sitting laden every single day and more data cap is being issued and distributed. In terms of notaries, we just wrapped our third election closing. And so we're at about 60 active notaries soon once they're all onboarded and willing to participate. Last year, we spent half of it at about 10 and half of it at about 20. And so we're three-axing the amount of notaries in the program and also like significantly increasing the amount of data cap available to them to serve. And this is mostly an experiment to see how the program actually scales. Right now, if you go and work with a notary, it takes about 10 days to allocate data cap. So it's very important to us that we work towards bringing that number down. So upcoming focus areas, well, actually these are more like current focus areas and then upcoming development. First is what I just mentioned, what we call TTD that stands for Time Due Data Cap. This is like our top line sort of success metric in the Filecoin Plus program. We wanna ensure the data cap is available and attainable in like the one hour or the three day range at all of these levels. And so that includes having more notaries, having the notaries be more active, improving the UX for clients, automation and scaling for data cap applications, et cetera. You can see the little flow chart on the right which sort of walks you through how we want to move towards more automation, reduce manual interaction points and only bring in the human interaction in the cases where things are starting to go awry. Focus on risk, reducing the amount of touch points, more focus on sort of on-chain tracking and the automation of like flags for like risky behavior or abuse or profiteering of the program. Like we're thinking about this as like us building out a reputation system or a credit system for data owners and search providers inside the Filecoin Plus world. And then generally evolving the no-reader a little bit to go from what we call like the gatekeeper-esque functionality to more like guardians as the network that changes the way in which they interact but of course comes with discussions around their incentives, how to retain their attentiveness and overall have them deliver a service effectively to the network while still maintaining the best interest to the network and not themselves and participate in governance in addition to just what the community does today. So how do you wanna get involved? We'd love to have you come join us in the Filecoin Plus channel in Slack. We have bi-weekly governance calls. You should come there. We're working on sort of improving our processes. We'd love to have you as, you know think about notary incentives, think about how we wanna evolve processes and system design. And generally as we move towards like structures that are similar to like a doubt we're spending our working groups experimenting with governance, system design, et cetera and generally overhauling the incentive and long-term ideation around how Filecoin Plus will continue to serve the network into the future. And we're having a Filecoin Plus half day summit at Austin Texas today before Phil Austin if you haven't heard of it yet. So Phil Austin will happen June 8th. We're doing Filecoin Plus June 7th. This will be a virtual first event. So even if you can't make it in person we'll be broadcasting this live streaming it of course, recording it. So definitely more information to follow on that keep your eyes out. Switching gears for the last couple of minutes here to talk about Slingshot. Slingshot is a community program for data processes and developers that awards the storage of real valuable and usable data on the Filecoin network. So this is completely different than Filecoin Plus but you'll see some similar threads and sort of enclosing hopefully I'll be able to share with you why I'm talking about both these programs and why they're compelling for you. So Slingshot overall has been running since Filecoin went main head which is just about 18 months to the day. Through these 18 months we've had about 37-ish petabytes of data on boarded the network comprising of open data sets. So there's like 1661 data sets all the way from like COVID data to environmental data, the governmental data and like random public use things that were found that could be interesting source from scientific organizations, non-profits NGOs, et cetera. We also have a data explorer which is pretty compelling in terms of helping you identify where you can retrieve that data from its mirror on Filecoin as opposed to needing to pay for an AWS retrieval service or needing to run your compute operations in different cloud provided services. Here you can actually get it effectively for free or extremely cheaply from the Filecoin network itself. Recently in the last six months sort of too couple with just this onboarding thing we've added two different programs as well that are focused on the longevity and the long-term viability that data being onboarded. So on the left you have Evergreen. The idea with Evergreen is effectively as Slingshot deals reach their end of life we want to continue renewing them. So right now this is a little bit of a manual we're very excited for the Filecoin virtual machine to show up so that we can automate some of these things but the idea here is tracking the pieces of data that are in expiring deals and then incentivizing the rebuilding of replicas for them to ensure longevity, permanence, long-term availability of that open data on the Filecoin network. The primary driving incentive here is actually data cap. And so this goes back to the first part of the conversation about how are we able to do this? The idea is that by adhering to an extremely high standard of quality storage and showing that we're getting to do things that push the definition of how to make data useful on Filecoin, how to contribute to the best practices around that both for the data owner side of the house but also for the search provider side of the house we're able to justify to the noted reason they should sort of support the running of this program. On the right we have recovery as an effort and within recovery specifically the Restore program Restore program specifically started out around an incident where we lost a little bit of data well a certain amount of data after like a freak accident effectively firing a data center and so we built this as like the first step towards like self-healing where we can design systems where data on screenshot if it's lost and this is specifically not in terms of end of life for deals but rather like a sector but false or a search provider decides to walk away. How do we rebuild and self-heal from that? Like how do we maintain a healthy number of live rather than any given point in time? How do we make sure they're available? Both of these programs are currently open and evergreen especially if you are thinking about a search writing operation and listening to this it's a pretty interesting one to get started definitely check it out. Why I'm telling you about this quick primer on DAO's I think you guys are all aware of this decentralized autonomous organizations typically referred to as a combination of community plus some sort of contract. What we care about or what I care about right now is specifically DAO's that care about data so the creation of value through shared data or data sets this is not just limited to the monetization of demand for data though that is one excellent use case and one way in which data does have often been described and arguably the data DAO sort of set that ideation process forward with that scope. So what I'm trying to structure right now is like using these components together it's a very complicated approach on the right I'm not gonna dig into it too much here but the idea is that we've got a component we've got a set of components that come together very nicely to build a system that can in a clean manner set up data for long-term preservation on the Filecoin network. So with just SlingShout you've got onboarding data with Evergreen you've got the aspect of guaranteeing its permanence with the store you can do the self-healing and the rebuilding and then we have this component that we've used in SlingShout for a while via the deal bot where we check for retrieval success rates. So we've got retrieval success rates as a services I like to call it where we do sampling and retrievability and have metrics on the health of the data itself so even though it might be alive and the deal might be good that doesn't actually mean that it's accessible and so by having a path to onboard, maintain and show that it's repaired and have a way in which it's like all those things happen in the way that the clients need access that data are able to. We do have all these nice components that sort of come together and can leverage like significant automation and development in the future to build like an interesting data structure. And so we'd love to have you involved if you're interested in this conversation lots of work to do ahead of us here definitely in terms of defining the incentives of governance, et cetera leveraging the virtual machine itself of course also just the data architecture and how we want to do the analysis how we want to like build that sort of reputation aspect that I mentioned. So specifically if you'd be interested in thinking about these things further please reach out to me either on Slack or Twitter or this QR code should actually take you straight to our definition and JD for one of the roles on the team. That doesn't have to be the role you want. I just wanted to put it up there as like a hey if you want to get involved sooner definitely check that out. Awesome. I know I'm a little bit behind schedule. Thank you all for bearing with me and getting on my speed train as you were hearing the time limit for this presentation. Thank you for having me. Alex in the accent. Congrats once again on the event. I'm looking forward to being there in person next time. In case there's any questions please feel free to chase me down. I don't know if we have the ability to do the questions right now because I know there's a pretty packed agenda to keep you guys going. I'm going to rely on the host to let me know. If not really appreciate your attention. Thanks a lot deep for that overview of Filecoin Plus and Slingshot and their various aspects. We have time for maybe one question to the presenter. Hi, yeah just really quick. So I understand storage reminders will want to participate in verified deals to have an increased percentage of the block rewards. What is the incentive for clients to want to get notarized or verified to have more data cap? Yeah, great question. So because storage providers have that incentive the amount of actual fill or like money that gets charged to make those deals is significantly lower. Actually the average verified deal in the network today is completely free. So the incentive for KYC and due diligence process is effectively you're trading that time, that effort, that loss of complete privacy on what you're doing and adding a little bit more transparency for getting absurdly subsidized deal making on the network. So especially as storage providers start to build long-term businesses on the network and want to offer the ability to store data at scale safely instead of having to pay in fill you get to effectively extend that data cap and then spend it. And so that's why it's extremely interesting and compelling for clients to go to that process. Of course, we are working on making that process easier and faster and all of that. And so that is also good, hopefully for clients in the long-term but that is the primary incentive today. I'd say there's a secondary social incentive as well where like doing this stuff in public means that everybody in the network sort of sees that you're trying to do the right thing sees the work that you're putting in behind it helps you sort of gain clout for long-term partnerships, long-term relationships hopefully favorable deal making climates in the future. But then coming into Falkland plus day for example we have a couple of people on both sides the marketplace that want to be presenting share the work that they've done the tooling they've done and how they're contributing to the success of the network. So I'd say, yeah, the monetary incentive have been extremely cheap deal the social incentive of building cloud in the network.