Matan Levi

342 posts

Matan Levi banner
Matan Levi

Matan Levi

@Matan5191

Sr. #AI Research Scientist @IBM | Taming LLMs | Ph.D. #CS @bengurionu

انضم Ocak 2022
338 يتبع51 المتابعون
تغريدة مثبتة
Matan Levi
Matan Levi@Matan5191·
We just released our preprint "CyberPal. AI: Empowering LLMs with Expert-Driven Cybersecurity Instructions." by @IBMResearch. Check out how our approach improves cybersecurity AI performance by up to 24% across a variety of tasks: arxiv.org/pdf/2408.09304… #IBM #LLMs 🧵 >>
Matan Levi tweet media
English
4
10
15
1.6K
Matan Levi أُعيد تغريده
clem 🤗
clem 🤗@ClementDelangue·
"But here is what we found when we tested: We took the specific vulnerabilities Anthropic showcases in their announcement, isolated the relevant code, and ran them through small, cheap, open-weights models. Those models recovered much of the same analysis. Eight out of eight models detected Mythos's flagship FreeBSD exploit, including one with only 3.6 billion active parameters costing $0.11 per million tokens. A 5.1B-active open model recovered the core chain of the 27-year-old OpenBSD bug." aisle.com/blog/ai-cybers…
English
110
345
2.4K
719.6K
Matan Levi أُعيد تغريده
המחנך הפיננסי
המחנך הפיננסי@FinancialEduX·
אחד הסרטונים האהובים עליי של ביל אקמן. הקרן שלו ירדה ביותר מ30% הוא נתבע על ידי משקיעים היה באמצע גירושין וקרן אקטיביסטית ניסתה להשתלט על פרשינג מה השיטה שלו לצאת ממצבים כאלו. ריבית דריבית עובדת בכל אספקט בחיים שלנו
עברית
17
34
729
49.3K
Matan Levi
Matan Levi@Matan5191·
I tested how vulnerable my ClawdBot (by @openclaw) is to indirect prompt injection (via email). It’s powerful — but if you connect it to inbox/WhatsApp/Telegram, you must harden it. Self-check your setup before someone else does. medium.com/p/one-email-is… #Clawdbot #MOLTBOT
English
0
0
0
92
Matan Levi
Matan Levi@Matan5191·
@Cyburgerim אני מניח שזה שילוב של מספר אינטרסים. אגב, לא מופרך שהם הריצו את אותה התקפה גם עם המודלים הפנימיים שלהם, ואז הם רצו לעשות benchmarking מול ה frontier models כמו קלוד כדי להבין איפה המודלים שלהם עומדים מבחינת יכולות.
עברית
0
0
1
13
Matan Levi
Matan Levi@Matan5191·
@Cyburgerim @Cyburgerim אלא אם זה היה ניסוי כלים מכוון כדי להבין איפה ה thresholds של מנגנוני הגילוי של אנתרופיק
עברית
3
0
2
238
Cyburger
Cyburger@Cyburgerim·
מחשבות לתחילת שבוע: הסיכוי שתקיפה במימון מדינתי, סיני או אחר, היתה נראית ככה - הוא נמוך מאוד.
Cyburger@Cyburgerim

חתיכת סיפור לסופ"ש! ✨ כנראה לראשונה: דיווח של Anthropic על תקיפה, ניישן-סטייט סיני לטענתם, שניצלה את המודלים שלה לבצע מתקפות agentic מלאות על ארגונים אחרים באופן כמעט אוטונומי 🤯 - דווח אתמול על מתקפה שבוצעה כבר לפני חודשיים - השימוש בAI ברובו היה פחות ל"התייעצות" ואיך לתקוף, אלא יותר בצד הagentic - משמע, *לבצע את התקיפות* - ניסיונות חדירה לכ30 ארגונים (טק, ממשל, פיננסים ועוד), הצליחו ב"מספר קטן של מקרים" - קבלת טרגט ממפעיל אנושי והרצת פריימוורק דרך Claude שעוקף את ההגנות שלהם באמצעות ג'יילברייק ופיצול כל שרשרת המתקפה למיני-משימות שכל אחת נראתה תמימה בפני עצמה - שתבינו את הטרלול מה קרה פה כמעט בצורה אוטונומית (הם מתארים אלפי בקשות בשניה): שימוש בכתיבת קוד תקיפה, הרצת קוד (אייג'נטס), שרשור פעולות תקיפה, חיפוש מידע משלים באינטרנט, מציאת חולשות, שימוש בcreds, שימוש בכלים אחרים דרך MCP, הרצת אקספלויטס, חילוץ מידע מהטרגט, קבלת החלטות כמעט אוטונומית (מי אמר Agentic AI CTF ולא קיבל??) ואפילו דוקומנטציה של התקיפה - מדי פעם היה צריך קבלת החלטה אנושית - כמו לדוגמא שנמצא DB רגיש, creds - המפעיל היה צריך לתת הנחיה איך להמשיך - טוענים ש80-90% בוצע ללא גורם אנושים (4-6 פעמים היו נדרשים בקבלת החלטות קריטיות), אבל הקטע שהכי הזיה בכל הסיפור הזה, זה שהיו שלבים בתקיפה שלא עבדו בגלל הזיות (hallucinations) של קלוד עצמו על credentials לא נכונים או על מידע "פרטי" שחולץ אבל כבר היה פאבליק - אלו הרסו את אוטונומיות התקיפה. לפעמים טוב שיש הזיות 🍄 שיהיה אחלה סופ"ש, עם/בלי הזיות, מה שעושה לכם טוב 🙌

עברית
8
1
17
6.6K
Matan Levi أُعيد تغريده
Rohan Paul
Rohan Paul@rohanpaul_ai·
New IBM paper builds small security expert language models that beat bigger ones on key threat tasks. The authors build SecKnowledge 2.0, a dataset with expert formats and grounded evidence. They fine tune CyberPal 2.0 models from 4B to 20B on that data. The models learn to answer fast for simple prompts and show steps for harder ones. Tests cover core threat knowledge and mapping bugs to the right weakness category. The 20B model ranks 1st on root cause mapping, and the 4B model is close behind. Average gains over their baselines are 7-14% across security benchmarks. Most gains come from stronger formats and evidence grounding rather than more compute. 8-bit and 4-bit versions keep most quality, which helps on prem deployments. The idea is that step-by-step, evidence-backed workflows let small models make reliable calls. ---- Paper – arxiv. org/abs/2510.14113 Paper Title: "Toward Cybersecurity-Expert Small Language Models"
Rohan Paul tweet media
English
5
3
16
3.8K
Matan Levi أُعيد تغريده
Andrej Karpathy
Andrej Karpathy@karpathy·
I quite like the new DeepSeek-OCR paper. It's a good OCR model (maybe a bit worse than dots), and yes data collection etc., but anyway it doesn't matter. The more interesting part for me (esp as a computer vision at heart who is temporarily masquerading as a natural language person) is whether pixels are better inputs to LLMs than text. Whether text tokens are wasteful and just terrible, at the input. Maybe it makes more sense that all inputs to LLMs should only ever be images. Even if you happen to have pure text input, maybe you'd prefer to render it and then feed that in: - more information compression (see paper) => shorter context windows, more efficiency - significantly more general information stream => not just text, but e.g. bold text, colored text, arbitrary images. - input can now be processed with bidirectional attention easily and as default, not autoregressive attention - a lot more powerful. - delete the tokenizer (at the input)!! I already ranted about how much I dislike the tokenizer. Tokenizers are ugly, separate, not end-to-end stage. It "imports" all the ugliness of Unicode, byte encodings, it inherits a lot of historical baggage, security/jailbreak risk (e.g. continuation bytes). It makes two characters that look identical to the eye look as two completely different tokens internally in the network. A smiling emoji looks like a weird token, not an... actual smiling face, pixels and all, and all the transfer learning that brings along. The tokenizer must go. OCR is just one of many useful vision -> text tasks. And text -> text tasks can be made to be vision ->text tasks. Not vice versa. So many the User message is images, but the decoder (the Assistant response) remains text. It's a lot less obvious how to output pixels realistically... or if you'd want to. Now I have to also fight the urge to side quest an image-input-only version of nanochat...
vLLM@vllm_project

🚀 DeepSeek-OCR — the new frontier of OCR from @deepseek_ai , exploring optical context compression for LLMs, is running blazingly fast on vLLM ⚡ (~2500 tokens/s on A100-40G) — powered by vllm==0.8.5 for day-0 model support. 🧠 Compresses visual contexts up to 20× while keeping 97% OCR accuracy at <10×. 📄 Outperforms GOT-OCR2.0 & MinerU2.0 on OmniDocBench using fewer vision tokens. 🤝 The vLLM team is working with DeepSeek to bring official DeepSeek-OCR support into the next vLLM release — making multimodal inference even faster and easier to scale. 🔗 github.com/deepseek-ai/De… #vLLM #DeepSeek #OCR #LLM #VisionAI #DeepLearning

English
561
1.6K
13.3K
3.3M
Matan Levi أُعيد تغريده
Brian Roemmele
Brian Roemmele@BrianRoemmele·
BOOOOOOOM! CHINA DEEPSEEK DOES IT AGAIN! An entire encyclopedia compressed into a single, high-resolution image! — A mind-blowing breakthrough. DeepSeek-OCR, unleashed an electrifying 3-billion-parameter vision-language model that obliterates the boundaries between text and vision with jaw-dropping optical compression! This isn’t just an OCR upgrade—it’s a seismic paradigm shift, on how machines perceive and conquer data. DeepSeek-OCR crushes long documents into vision tokens with a staggering 97% decoding precision at a 10x compression ratio! That’s thousands of textual tokens distilled into a mere 100 vision tokens per page, outmuscling GOT-OCR2.0 (256 tokens) and MinerU2.0 (6,000 tokens) by up to 60x fewer tokens on the OmniDocBench. It’s like compressing an entire encyclopedia into a single, high-definition snapshot—mind-boggling efficiency at its peak! At the core of this insanity is the DeepEncoder, a turbocharged fusion of the SAM (Segment Anything Model) and CLIP (Contrastive Language–Image Pretraining) backbones, supercharged by a 16x convolutional compressor. This maintains high-resolution perception while slashing activation memory, transforming thousands of image patches into a lean 100-200 vision tokens. Get ready for the multi-resolution "Gundam" mode—scaling from 512x512 to a monstrous 1280x1280 pixels! It blends local tiles with a global view, tackling invoices, blueprints, and newspapers with zero retraining. It’s a shape-shifting computational marvel, mirroring the human eye’s dynamic focus with pixel-perfect precision! The training data? Supplied by the Chinese government for free and not available to any US company. You understand now why I have said the US needs a Manhattan Project for AI training data? Do you hear me now? Oh still no? I’ll continue. Over 30 million PDF pages across 100 languages, spiked with 10 million natural scene OCR samples, 10 million charts, 5 million chemical formulas, and 1 million geometry problems!. This model doesn’t just read—it devours scientific diagrams and equations, turning raw data into a multidimensional knowledge. Throughput? Prepare to be floored—over 200,000 pages per day on a single NVIDIA A100 GPU! This scalability is a game-changer, turning LLM data generation into a firehose of innovation, democratizing access to terabytes of insight for every AI pioneer out there. This optical compression is the holy grail for LLM long-context woes. Imagine a million-token document shrunk into a 100,000-token visual map—DeepSeek-OCR reimagines context as a perceptual playground, paving the way for a GPT-5 that processes documents like a supercharged visual cortex! The two-stage architecture is pure engineering poetry: DeepEncoder generates tokens, while a Mixture-of-Experts decoder spits out structured Markdown with multilingual flair. It’s a universal translator for the visual-textual multiverse, optimized for global domination! Benchmarks? DeepSeek-OCR obliterates GOT-OCR2.0 and MinerU2.0, holding 60% accuracy at 20x compression! This opens a portal to applications once thought impossible—pushing the boundaries of computational physics into uncharted territory! Live document analysis, streaming OCR for accessibility, and real-time translation with visual context are now economically viable, thanks to this compression breakthrough. It’s a real-time revolution, ready to transform our digital ecosystem! This paper is a blueprint for the future—proving text can be visually compressed 10x for long-term memory and reasoning. It’s a clarion call for a new AI era where perception trumps text, and models like GPT-5 see documents in a single, glorious glance. I am experimenting with this now on 1870-1970 offline data that I have digitalized. But be ready for a revolution! More soon. [1] github.com/deepseek-ai/De…
Brian Roemmele tweet media
English
341
1.4K
7.5K
1.8M
Matan Levi أُعيد تغريده
Ilya Sutskever
Ilya Sutskever@ilyasut·
truly the greatest day ever🎗️
English
832
685
16K
1.8M
Matan Levi أُعيد تغريده
Aniket Didolkar
Aniket Didolkar@Aniket_d98·
🚨Reasoning LLMs are e̵f̵f̵e̵c̵t̵i̵v̵e̵ ̵y̵e̵t̵ inefficient! Large language models (LLMs) now solve multi-step problems by emitting extended chains of thought. During the process, they often re-derive the same intermediate steps across problems, inflating token usage and latency. Metacognitive Reuse: turn recurring LLM reasoning into concise, reusable “behaviors”. The model learns named skills from its own chains-of-thought and reuses them to think faster & cheaper. Arxiv 🔗 - arxiv.org/abs/2509.13237
English
4
35
209
45.2K
Matan Levi أُعيد تغريده
Aran Komatsuzaki
Aran Komatsuzaki@arankomatsuzaki·
RL’s Razor: On-policy RL forgets less than SFT. Even at matched accuracy, RL shows less catastrophic forgetting Key factor: RL’s on-policy updates bias toward KL-minimal solutions Theory + LLM & toy experiments confirm RL stays closer to base model
Aran Komatsuzaki tweet media
English
9
102
632
111.5K
Chris Koch
Chris Koch@chrisk99999·
@Matan5191 @OpenAI @Eric_Wallace_ Gotchya, ya the anti-refusal will probably be slightly higher capability than the jailbroken version. I think the anti-refusal training is pretty cheap, it can probably be done in a few H100 days (I'm not sure though).
English
1
0
1
16
Matan Levi
Matan Levi@Matan5191·
1/11 You don’t need a million-dollar budget to dent guardrails. Below I show how a single A100-80GB and pocket change can break @OpenAI 's #GPT OSS refusals mechanism. Here’s why—and what OpenAI’s risk paper does (and doesn’t) cover 👇
English
1
0
5
292
Matan Levi
Matan Levi@Matan5191·
@chrisk99999 @OpenAI @Eric_Wallace_ Yes, totally I agree that the anti-refusal is indeed the lower bound for adversaries *with training budget*. I thought it will be interesting to see what will be the lower bound if the adversary does not have a large training budget.
English
1
0
1
35
Matan Levi
Matan Levi@Matan5191·
@chrisk99999 @OpenAI @Eric_Wallace_ Since these kind of attacks are the most easy and most cost effective to perform, it can serve as some kind of a lower bound (minimum gain) a malicious actor can gain from the open source model, which IMO as interesting as the case of increasing the ceiling.
English
2
0
0
34