DeAI: decentralized artificial intelligence
  • Introduction
    • General Terminology
  • Landscape
    • Data Providers
    • Computing Power
    • Model Training Task
    • Challenges
  • Privacy Preservation
    • Data Process
    • Privacy Preserved Training
    • Federated Learning
    • Cryptographic Computation
      • Homomorphic encryption
      • Multi-Party Computation
      • Trusted Execution Environment
    • Challenges
  • Security
    • Data Poisoning
    • Model Poisoning
    • Sybil Attacks
    • Impact of Large Models
    • Responsibility
  • Incentive mechanisms
    • Problem Formulation
    • Contribution Evaluation
    • Copyright
  • Verification of Computation
    • Computation on Smart Contract
    • Zero-Knowledge Proof
    • Blockchain Audit
    • Consensus Protocol
  • Network Scalability
    • Local Updating
    • Cryptography Protocol
    • Distribution Topology
    • Compression
    • Parameter-Efficient Fine Tuning
  • Conclusion
Powered by GitBook
On this page
  1. Landscape

Data Providers

Current LLMs (e.g., GPT-4, Gemini, Llama) are trained with data of trillions of tokens, indicating the imminent depletion of high-quality data. To enhance model capability, incorporating more high-quality private data is imperative. From the perspective of data providers, concerns arise regarding privacy disclosure via data sharing, as data is utilized by the model training task and executed on computing powers. Importantly, models trained on their data should not disclose privacy, such as personally identifiable information (PII). Furthermore, incentivizing mechanisms are crucial to encourage data providers to produce more high-quality data. However, the current model training process lacks incentivization, failing to reward high-quality data via model monetization.

PreviousLandscapeNextComputing Power

Last updated 11 months ago