DeAI: decentralized artificial intelligence
  • Introduction
    • General Terminology
  • Landscape
    • Data Providers
    • Computing Power
    • Model Training Task
    • Challenges
  • Privacy Preservation
    • Data Process
    • Privacy Preserved Training
    • Federated Learning
    • Cryptographic Computation
      • Homomorphic encryption
      • Multi-Party Computation
      • Trusted Execution Environment
    • Challenges
  • Security
    • Data Poisoning
    • Model Poisoning
    • Sybil Attacks
    • Impact of Large Models
    • Responsibility
  • Incentive mechanisms
    • Problem Formulation
    • Contribution Evaluation
    • Copyright
  • Verification of Computation
    • Computation on Smart Contract
    • Zero-Knowledge Proof
    • Blockchain Audit
    • Consensus Protocol
  • Network Scalability
    • Local Updating
    • Cryptography Protocol
    • Distribution Topology
    • Compression
    • Parameter-Efficient Fine Tuning
  • Conclusion
Powered by GitBook
On this page
  • Differential privacy
  • Privacy Regularization
  1. Privacy Preservation

Privacy Preserved Training

Differential privacy

Differential privacy, defined as ensuring that adjacent data cannot be distinguished, offers privacy protection through an information-theoretic guarantee. In practice, differential privacy is implemented by adding noise to data, gradients, output, or objective functions. However, while differential privacy is crucial for preserving privacy, it may blur long-tail examples in data distributions, resulting in reduced accuracy, particularly for underrepresented groups. Despite its potential negative impact on pretrained model performance, differential privacy in fine-tune tasks can maintain model utility.

Privacy Regularization

Privacy regularization introduces penalties for generating privacy-sensitive information. For instance, PPLM introduces instruction tuning with Direct Preference Optimization (DPO) to reward generations that distinguish between publicly shareable and privacy-sensitive information.

PreviousData ProcessNextFederated Learning

Last updated 11 months ago