Francesco Pinto @ Neurips 2024

43 posts

Francesco Pinto @ Neurips 2024 banner
Francesco Pinto @ Neurips 2024

Francesco Pinto @ Neurips 2024

@FraPintoML

Postdoc #UChicago, ex-#UniversityOfOxford, #Meta,#Google, #FiveAI, #ETHZurich Trustworthy and Privacy-Preserving ML Email: [email protected]

Oxford, UK Katılım Ocak 2022
176 Takip Edilen54 Takipçiler
Francesco Pinto @ Neurips 2024 retweetledi
Zhaorun Chen
Zhaorun Chen@ZRChen_AISafety·
🚀 Introducing 𝐒𝐚𝐟𝐞𝐖𝐚𝐭𝐜𝐡! 🚀 While generative models 👾🎥 like Sora and Veo 2 have shown us some stunning videos recently, they also make it easier to produce harmful content (sexual🔞, violent🙅‍♂️, deepfakes🧟‍♂️). 🔥 𝐒𝐚𝐟𝐞𝐖𝐚𝐭𝐜𝐡 is here to help 😎: the first MLLM-based video guardrail model designed to follow customized safety policies and provide guardrails with precise explanations in a zero-shot manner. In addition, we also introduce SafeWatch-Bench📊, a 2M+ high-quality video guardrail dataset covering over 30 unsafe video scenarios from various real-world platforms and SOTA generative models to comprehensively cover all potential risks. 🧐Why SafeWatch? 👉1. Strong policy-following: trained on diverse videos and policy taxonomies, yielding high generalizability to unseen scenarios and subtle policy definitions. 👉2. High Inference Speed: introducing two plug-and-play modules to process policies in parallel and prune irrelevant video tokens, reducing inference costs and eliminating positional bias. 👉3. In-depth explanations: trained on high-quality explanations from SafeWatch-Bench📊 labeled by a rigorous multi-agent consensus pipeline and verified by human experts. We evaluate SafeWatch on a large variety of guardrail tasks: 1️⃣ On both real-world and generative video subsets of SafeWatch-Bench, SafeWatch outperforms SOTAs, including GPT-4o, by 29.2% and 27.2% on average, while requiring much less inference time. 2️⃣ On 5 existing video guardrail benchmarks, SafeWatch achieves 87.1% accuracy, consistently outperforming previous SOTAs. 3️⃣ On 4 new video categories and unseen policy taxonomies, as well as 4 different prompting tasks, SafeWatch maintains high accuracy and outperforms GPT-4o (renowned for its zero-shot generalizability). 🔥🔥 Our project has been released: 👉Paper link: arxiv.org/pdf/2412.06878 👉Project page: safewatch-aiguard.github.io 👉Code (coming soon): github.com/BillChan226/Sa…
Zhaorun Chen tweet media
English
4
15
40
5.1K
Francesco Pinto @ Neurips 2024
I’ll be at #NeurIPS2024 from now to Sunday. DM here or on Whova to have a chat about (multimodal) large language models privacy, memorisation, training strategies using synthetic data, agents, judges, distribution shift robustness, hallucinations and uncertainty estimation.
English
0
0
3
92
Francesco Pinto @ Neurips 2024
🧵 [3/3] Special thanks to all coauthors: Adam Davies, Ashkan Khakzar, Anjun Hu, Arshia Hemmat, Jianhao Yuan, Tom Lamb, Jiyang Guan, Philip Torr. Work done at @OxfordTVG
English
0
0
1
67
Francesco Pinto @ Neurips 2024
(2/3)🔎 These models may regurgitate names, addresses, card numbers, ids. 🧑‍🔬 We find high input training resolution and stronger pre-training can significantly reduce the chances of regurgitation.
Francesco Pinto @ Neurips 2024 tweet media
English
1
0
0
93
Francesco Pinto @ Neurips 2024
(1/3)🔥Multi-Modal LLMs (MLLMs) can respond to questions about document scans. How safe are they? Come at Hall C #2300 1.30pm to find out! 🧠Attackers may successfully query MLLMs to extract Personally Identifying Information! 🚨 arxiv.org/abs/2407.08707
Francesco Pinto @ Neurips 2024 tweet media
English
1
2
6
517
Francesco Pinto @ Neurips 2024
3/3🤖 Simple prompting and editing outperform traditional augmentations, producing more robust models with fewer augmented samples. 🛑 Given the quality of generative models, filtering is no longer required to attain improved performance. @OxfordTVG @DYDYYDYYYD @adamdaviesnlp
Francesco Pinto @ Neurips 2024 tweet media
English
0
0
1
99
Francesco Pinto @ Neurips 2024
2/3,⌨️ Prompting Text-to-Image generators proves to be an extremely effective (SOTA) and interpretable approach to synthesize interventional data for augmentation. 📈We extensively study the impact on robustness of conditioning mechanisms, prompting strategies and filtering.
Francesco Pinto @ Neurips 2024 tweet media
English
1
0
0
87
Francesco Pinto @ Neurips 2024
1/3,🧪🤖 What's the best way to improve model robustness to distribution shift using synthetic data? 💪 Come to Hall C 4-9 #912 #ICML2024 to find out! 💥Classifiers fail to recognise objects observed in previously unseen settings. 🧪 Can #StableDiffusion be used to fix this?
Francesco Pinto @ Neurips 2024 tweet media
English
6
3
7
1.1K
Francesco Pinto @ Neurips 2024
[2/2] "Strong Copyright Protection for Language Models via Adaptive Model Fusion" tinyurl.com/yc8yskkb GenLaw and Foundation Models in the Wild workshops 👋 Let's grab a coffee and chat about uncertainty, privacy, memorization, robustness, synthetic data, multimodal agents
English
1
0
0
82
Francesco Pinto @ Neurips 2024 retweetledi
Francesco Pinto @ Neurips 2024
🔥 Excited to be co-organizing this #ECCV2024 workshop with an outstanding line-up of speakers! 🗣️ 🔎Submit if you got papers with new benchmarks and analyses inspecting Emergent Visual abilities ✔️ or limitations ❌of Foundation Models! 🤖
Oxford Torr Vision Group@OxfordTVG

🔥 #ECCV2024 Showcase your research on the Analysis and Evaluation of emerging VISUAL abilities and limits of foundation models 🔎🤖👁️ at the EVAL-FoMo workshop 🧠🚀✨ 🔗 sites.google.com/view/eval-fomo… @phillip_isola @sainingxie @chrirupp @OxfordTVG @berkeley_ai @MIT_CSAIL

English
1
1
7
761