Conversation
Create DataDEX.md
add license info
Noc2
left a comment
There was a problem hiding this comment.
Thanks for the application. In general, I recommend to provide more technical details (functions, programming language, etc.) and to focus on the unique parts of your application. For example, we already have quite a few AMM based exchanges in our ecosystem, so we probably don’t fund just another AMM at this stage. But how do you for example proof that if you buy data, you actually receive the correct data? Could potentially provide an example? How does your solution compare to the ocean protocol?
Thanks for your comment. I added a data registration and checking design in the application to ensure the basically security and correctness of data reading. pls check it. We currently develop in Java, and be preparing refactor essential modules in Rust. There are many differences compare to ocean protocol:
|
Noc2
left a comment
There was a problem hiding this comment.
Thanks for the additional information. Another issue I see is how do you prevent people from “stealing” data. For example, I buy someone else's data, I change a very small part of it in a way that the hash of data changes and I just sell the same data? Maybe you could explain the DataGraph a little bit more. Additionally how do you ensure the availability of the data. For example, I buy data on the exchange, but it’s actually no longer available and I can not download it.
Thanks for comment. DataGraph is designed with multi-properties schema. After register the data, the Data Owner could be verified the ownership from multiple dimensions, not only the userid. If the malicious data owner just modifies a small part, other dimensions will conflict to the real data, the registration cannot be successfully submitted and cannot be successfully verified. We will also reward community members who voluntarily find malicious data or errors (we hope that Dataset Maker will play a leading role), based on data characteristics analysis and relevant in data technology. and then owners with data faults should be punished. We have rich experience in anti fake products for e-commerce platforms with this similar technology. If the data owner could not guarantee the data availability, including the data non-exists or the data unregistered etc. The tasks scheduled by the computing node(Phala, Alita network etc.) to the device of data owner will fail, DataDEX can not check the verification of task, so the DataOwner can not get the rewards. |
Noc2
left a comment
There was a problem hiding this comment.
Okay thanks for the reply. Could you integrate this into the application, ideally even into the deliverables? We have quite some interest in supporting the development and research of on-chain mechanisms to identify “malicious data”, e.g. via a DAO or scheme to deal with data availability issues. In general, it would be helpful if the deliverables are more specific. For example, in which programming languages are you planning to develop the smart contract, front-end, app, etc. What are the functionalities of these deliverables? Btw. feel free to remove the deployment on roco, since this isn’t something necessarily helpful for others.
add DAO and DataGraph example
add development specifics
Thanks for your advice. I have integrated specifics in application, pls check it. |
|
Thanks for the update. I have a few follow-up questions: Is this your previous project: https://github.com/ALITANetwork? Why do you want to start using substrate now? Are the UIs part of the repo the ones, which you plan to develop? And are you familiar with burstcoin? |
Yes, AlitaNetwork is our previous project, which is a privacy-preserving edge computing network. One of the component accounting networks is developed based on burst. But afterwards, we met the situation that we have too much work in the development of the public chain. and found that the openness and friendship of substrate can greatly help us reduce the workload and invest our time in data and computing that we are more familiar with. In addition, polka has rich ecological resources, and the staking economic model is more fitted for us to expand the community, so we plan to migrate to substrate. Since the new project will be developed based on substrate, we hope that the previous UIs need to be totally refactored. Thanks. |
|
Hi @Noc2 , thanks for your time. It's a nice conversation. |
|
Sorry for the late reply. So the UI will be based on your previous work, correct? I would still be interested in learning more about your relationship with burstcoin. |
Thanks for your reply.
Additionally, we have no cooperation with the burstcoin project or the community. |
|
Hi @Noc2 What are your main concerns about the previous project? I would be very glad to hear from you suggestions. Thanks. |
|
Thanks for the additional information. In this case, I recommend to make sure that you properly point out in your repos which work was done by you and which was copied/cloned by burstcoin. I have another question: Are you planning to deliver: “Data Registeration Entry'' and “Data Graph DAO” as part of the grant? In the meantime I will share the application with the rest of the team. |
|
Hi @gravity-link, a few questions from my side:
|
revise deliveries milestone.
Thanks for the suggestions. |
Thanks for the comment.
|
Thanks for the update. I’m happy to go ahead with it, once you have updated the repos. Feel free to ping me here. One additional suggestion from my side: It might make sense to split the deliveries into two milestones. We only pay you once you successfully delivered a milestone and currently you would need to deliver a lot before you get paid. |
split milestone
Hi @Noc2 Thanks for your advice. I have split the deliveries into two milestones. |
|
Hi @gravity-link, thanks for your reply. I have decided not to approve your application because (a) I think creating a new token contract for every dataset being sold seems inefficient, (b) linking a dataset to a limited amount of tokens sounds like artificial shortage, and (c) if I were to sell my personal data to the highest bidder, I would at least like to know who I'm selling to. I'd rather not have them bound to a token that might get passed around. I also wouldn't want to encourage people to sell their personal data, although I do appreciate the idea of at least making it a transparent marketplace! |
Hi @semuelle I am very appreciated that you gave us such an in-depth and professional review opinions of the application. a) Creating a token for each Dataset is mainly due to data pricing issue. Because the value of the data set is related to the data processing algorithm, scenario, sales scale, and even region and crowd, Dataset Maker will have capability to operate its own data supply and demand market. In our experience, if data pricing is too unified, the data market transaction volume would be very limited. b) I would like to further explain that limited tokens do not mean limited data volume. Dataset Maker can continuously invite new Data Owners to contribute data(permissions), but according to the price of Token, the incentives for later data Owners are different. Therefore, the limited Token actually provides a control capability of token circulation for the Dataset Maker. c) We not only hope that the Data Owner can gain data revenue, but also can trace the data usage, including who uses the data and how to use the data. Additionally, What tokenized is data readable privileges, not raw data. I think the mentioned above would be why the personal data should be tokenized. As we know that have been already a fact that Internet providers used personal data on a large scale. We just hope to help Data Owners get back their benefits they lost. I know that there is a huge difference between the Economic Model for Personal Data and Enterprise Confidential Data, and also hardworking on that, but I still believe that this is a matter of very social significance and commercial value, and I believe that we are one of the most experienced team to push it forward in progress. So if possible, please consider the feasibility of this application again, thank you very much! BTW. We will upload the revised english version of Light-Paper tomorrow, in which will add specifics on pricing etc. and will post it here ASAP, please refer to it. Download here please. |
|
Hi @gravity-link. After some consideration, I'm sorry to inform you that I will not support this proposal either due to the following reasons:
|
add reference
|
Thanks for your honest and professional opinions. I would like to explain following:
In addition, cost calculation and data price measurement are already common pricing strategies. I added reference about AWS data exchange pricing in the bottom of application.
Awsome, it seems be a privacy computing project. But I am not sure why it aims on personal data, and it seems to be suitable for non-personal data too. I hope it would be a compute node of DataDEX. I will follow its progress and would like to learn from its scheme.
I will improve this.
DataDEX designed for high quality and high volume of dataset, not only the number of dataset. I have added a reference that Google built one high quality dataset distributed to millions users mobile phones for Next Word model training.So I believe creating a token for a dataset of great value would be make sense.
There are 3~5 engineers working for Alita project. Some components have not open source but soon.
Thanks for your advice. I will continuously add specifics in application. |
add Task Oracle specifics
|
Thank you for your clarifications @gravity-link. Your first point in specific is valid and I had misunderstood this aspect of your application. However, I don't see how any of your updates or other answers address my concerns and I stand by my decision. Others might be convinced though, let's give the application a bit more time. |
add Scenario and external reference URLs
|
Hi @alxs , @semuelle , |
|
Closing since there seems to be no interest from other members of the committee either. As previously stated, there are a number of issues you haven't addressed yet and the proposal lacks technical details, neither of which has been improved in your last update. Feel free to reopen an application when you're further along with your project and are able to address these issues. |
Open Square Network - BlockChain Based Crowdsourcing and Reputation Platform
Grant Application Checklist