Skip to content

Commit 5a7f39e

Browse files
committed
Outline and extended description
1 parent e85ad30 commit 5a7f39e

1 file changed

Lines changed: 27 additions & 16 deletions

File tree

README.md

Lines changed: 27 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -13,28 +13,41 @@ allowing to train machine learning models with full privacy guarantees. In this
1313
several methods for privacy-preserving data analysis, and how these techniques can be used to safely train
1414
ML models _without_ actually seeing the data.
1515

16+
### Description
17+
18+
Privacy guarantees are **the** most crucial requirement when it comes to analyse sensitive data. These requirements could be sometimes very stringent, so that it becomes a real barrier for the entire pipeline. Reasons for this are manifold, and involve the fact that data could not be _shared_ nor moved from their silos of resident, let alone analysed in their _raw_ form. As a result, _data anonymisation techniques_ are sometimes used to generate a sanitised version of the original data. However, these techniques alone are not enough to guarantee that privacy will be completely preserved. Moreover, the _memoisation_ effect of Deep learning models could be maliciously exploited to _attack_ the models, and _reconstruct_ sensitive information about samples used in training, even if these information were not originally provided.
19+
20+
*Privacy-preserving machine learning* (PPML) methods hold the promise to overcome all those issues, allowing to train machine learning models with full privacy guarantees.
21+
22+
This workshop will be mainly organised in **three** main parts. In the first part, we will introduce the main concepts of **differential privacy**: what is it, and how this method differs from more classical _anonymisation_ techniques (e.g. `k-anonymity`). In the second part, we will focus on Machine learning experiments. We will start by demonstrating how DL models could be exploited (i.e. _inference attack_ ) to reconstruct original data solely analysing models predictions; and then we will explore how **differential privacy** can help us protecting the privacy of our model, with _minimum disruption_ to the original pipeline. Finally, we will conclude the tutorial considering more complex ML scenarios to train Deep learning networks on encrypted data, with specialised _distributed federated_ _learning_ strategies.
23+
1624
### Outline
1725

18-
The tutorial is organised in four parts (more or less 1h each):
26+
- **Introduction**: Brief Intro to `PPML` and to the workshop (`10 mins`) [SLIDES](https://speakerdeck.com/leriomaggio/ppml-scipy)
1927

20-
- Introduction: overview and objectives
21-
1. Sensitive Data and Anonymisation techniques:
22-
- quasi-identifiers
23-
- k-anonimity
28+
- **Part 1**: Programming Privacy (`90 mins`)
29+
- De-identification
30+
- K-anonimity and limitations
31+
- Differential Privacy
32+
- Intro to Differential Privacy for Machine Learning
2433

25-
2. Introduction to Differential Privacy
34+
- Break (`10 mins`)
2635

27-
3. ML Model attacks
28-
- Adversarial and FSGM
29-
- Membership Inference Attack
30-
- Differential Privacy for ML Models
36+
- **Part 2**: Strengthening Deep Neural Networks (`60 mins`)
37+
- ML Model vulnerabilities: Adversarial Examples and _inference attack_
38+
- DL training with Differential Privacy
3139

32-
4. Federated Learning
33-
- Intro to FL
34-
- Homomorphic Encryption & FL
40+
- **Break** (`5 mins`)
3541

42+
- **Part 3**: Primer on Privacy-Preserving Machine Learning (`60 mins`)
43+
- DL training on (Homomorphically) Encrypted Data
44+
- Federated Learning
3645

37-
#### Notebooks:
46+
- **Closing Remarks** (`5 mins`)
47+
48+
#### Notebooks
49+
50+
Quick access to each notebooks, also to open on **Anaconda Notebooks**
3851

3952
1 Data Anonimisation
4053
- References: [![open_in_anaconda](https://static.anaconda.cloud/content/a22d04e8445b700f28937ab3231b8cded505d0395c63b7a269696722196d5415)](https://anaconda.cloud/api/nbserve/launch_notebook?nb_url=https%3A%2F%2Fraw.githubusercontent.com%2Fleriomaggio%2Fppml-tutorial%2Fmain%2F1-data-anonimisation%2F1-references.ipynb)
@@ -82,8 +95,6 @@ The tutorial is organised in four parts (more or less 1h each):
8295

8396
- Flower FL [![open_in_anaconda](https://static.anaconda.cloud/content/a22d04e8445b700f28937ab3231b8cded505d0395c63b7a269696722196d5415)](https://anaconda.cloud/api/nbserve/launch_notebook?nb_url=https%3A%2F%2Fraw.githubusercontent.com%2Fleriomaggio%2Fppml-tutorial%2Fmain%2F4-federated-learning%2F4-flower-tutorial-pytorch.ipynb)
8497

85-
86-
8798
## Get the material
8899

89100
Clone the current repository by running the following instructions:

0 commit comments

Comments
 (0)