Michael Anthony's picture
Building on HF

Michael Anthony PRO

MikeDoes

AI & ML interests

Privacy, Large Language Model, Explainable

Recent Activity

reacted to their post with ๐Ÿ”ฅ about 19 hours ago
Are you sure the open-source model you just downloaded is safe? A recent paper on "Privacy Backdoors" reports a new vulnerability where pre-trained models can be poisoned before fine-tuning them. This is a serious challenge for everyone building on open-source AI. Instead of just pointing out problems, we believe in finding better solutions. To understand this threat, the researchers needed to test their attack on realistic data structures. They needed a dataset that could effectively simulate a high-stakes privacy attack, and we're proud that our Ai4Privacy dataset was used to provide this crucial benchmark. The paper reports that for our complex dataset, the privacy leakage on a non-poisoned model was almost zero. After the backdoor attack, that number reportedly jumped to 87%. Ai4Privacy dataset provided a realistic benchmark for their research. Our dataset, composed of synthetic identities, helped them demonstrate how a poisoned model could dramatically amplify privacy leakage. This is why we champion open source: it enables the community to identify these issues and develop better, safer solutions together. Kudos to the authors Yuxin Wen, Leo Marchyok, Sanghyun Hong, Jonas Geiping, Tom Goldstein, and Nicholas Carlini, University of Maryland and Google DeepMind. ๐Ÿ”— Read the research to understand this new challenge: https://arxiv.org/pdf/2404.01231 ๐Ÿš€ Stay updated on the latest in privacy-preserving AIโ€”follow us on LinkedIn: https://www.linkedin.com/company/ai4privacy/posts/
posted an update 1 day ago
Are you sure the open-source model you just downloaded is safe? A recent paper on "Privacy Backdoors" reports a new vulnerability where pre-trained models can be poisoned before fine-tuning them. This is a serious challenge for everyone building on open-source AI. Instead of just pointing out problems, we believe in finding better solutions. To understand this threat, the researchers needed to test their attack on realistic data structures. They needed a dataset that could effectively simulate a high-stakes privacy attack, and we're proud that our Ai4Privacy dataset was used to provide this crucial benchmark. The paper reports that for our complex dataset, the privacy leakage on a non-poisoned model was almost zero. After the backdoor attack, that number reportedly jumped to 87%. Ai4Privacy dataset provided a realistic benchmark for their research. Our dataset, composed of synthetic identities, helped them demonstrate how a poisoned model could dramatically amplify privacy leakage. This is why we champion open source: it enables the community to identify these issues and develop better, safer solutions together. Kudos to the authors Yuxin Wen, Leo Marchyok, Sanghyun Hong, Jonas Geiping, Tom Goldstein, and Nicholas Carlini, University of Maryland and Google DeepMind. ๐Ÿ”— Read the research to understand this new challenge: https://arxiv.org/pdf/2404.01231 ๐Ÿš€ Stay updated on the latest in privacy-preserving AIโ€”follow us on LinkedIn: https://www.linkedin.com/company/ai4privacy/posts/
View all activity

Organizations

Ai4Privacy's profile picture Social Post Explorers's profile picture Mistral AI Game Jam's profile picture AI STATUS CODES's profile picture Swiss {ai} Weeks's profile picture VNCYBERS., JSC's profile picture