

Punya Syon Pandey
19 posts









Excited to have 3 accepted papers & 9 members of our @JinesisLab at #IASEAI2026, held at UNESCO, Paris🇫🇷! We reveal hidden authoritarian biases in #LLMs, and that fine-tuning can quietly erode model safety, exploring the risks we don't always see in AI 🔍🛡️ 🧵👇






Introducing our MTI-LLM @ NeurIPS work: CORE, the first metric to measure multi-agent interaction quality and quantify linguistic diversity in LLM interactions. Excellent work by our author team, @psyonp, @_yongjinny, @Jiarui_Liu_, @ZhijingJin. Paper: arxiv.org/abs/2508.11915














A quick look into DeepSeek’s safety guard: We find DeepSeek’s Llama Distill is >2x⚠️ as vulnerable to jailbreaking attacks as the original Llama. Seems to be a large safety risk. Stay tuned for our upcoming work @psyonp @SimkoSamuel @ZhijingJin

