Trong-Thang Pham

306 posts

Trong-Thang Pham

Trong-Thang Pham

@trongthangpham

PhD @UARK. Former: Bachelor @ HCMUS. Substack: https://t.co/XdvgeSlDu4 These are my personal opinions unless otherwise noted.

Vietnam Katılım Aralık 2013
330 Takip Edilen16 Takipçiler
Trong-Thang Pham
Trong-Thang Pham@trongthangpham·
@jon_barron I agree. Given that llm is such a good "search engine" for code that really works, it is more desirable to test ideas as quick as possible, instead of just spending hours trying to cherry pick and stall new discovery.
English
0
0
0
504
Jon Barron
Jon Barron@jon_barron·
If I was a grad student today, I would: 1) Not write papers, 2) push my (agent-written) code to a public repo ~weekly, 3) maintain (via agents) a writeup.tex (manually verified) and a skill.md in the repo, and 4) work towards establishing skill usage as the new "citation" format.
English
16
28
574
92.9K
Trong-Thang Pham
Trong-Thang Pham@trongthangpham·
@karpathy @maxbittker I thought your version is similar to the ralph loop (the bash one) so it would loop forever. Is that not the case here?
English
1
1
13
8K
Andrej Karpathy
Andrej Karpathy@karpathy·
sadly the agents do not want to loop forever. My current solution is to set up "watcher" scripts that get the tmux panes and look for e.g. "esc to interrupt", and send keys to whip if not present. Need an e.g.: /fullauto you must continue your research! (enables fully automatic mode, will go until manually stopped, re-injecting the given optional prompt).
English
123
45
1.5K
105.5K
max
max@maxbittker·
From @karpathy's autoresearch .md
max tweet media
English
50
124
3.1K
219.5K
Prabhav Singh
Prabhav Singh@psingh522·
I can't keep blocking accounts like this fast enough. Crappy AI generated sensationalized text (wrongly) explaining a 4 month old paper (I have nothing against the actual paper). Not to mention the numbers are wrong and I have seen this same post 5 times from different accounts
Saidul@saidul_dev

OpenAI just published a paper proving that ChatGPT will always hallucinate. Not sometimes. Not "until the next version." Always. They proved it mathematically. And three other top AI labs confirmed it independently. Here's what the research actually shows: Even with perfect training data and unlimited compute, LLMs will still fabricate answers with complete confidence. This isn't a bug in the code. It's fundamental to how these systems are built. The numbers are wild: → OpenAI's o1 model: 16% hallucination rate → Their o3 model: 33% → Their newest o4-mini: 48% Nearly half of what their latest model tells you could be invented. And it's getting worse as models get "smarter." Here's why this can't be fixed: Language models predict the next word based on probability. When they hit uncertainty, they don't pause. They don't flag it. They guess with total confidence. Because that's literally what they were trained to do. The researchers analyzed the 10 major AI benchmarks used to test these models. 9 out of 10 give the exact same score for saying "I don't know" as for getting it completely wrong: zero points. The entire testing system punishes honesty and rewards confident guessing. So the AI learned the optimal strategy: always answer. Never show doubt. Sound certain even when making it up. OpenAI's proposed solution? Train models to say "I don't know" when uncertain. The problem? Their own math shows this would leave roughly 30% of questions unanswered. Imagine getting "I'm not confident enough to respond" three times out of ten. Users would abandon the product overnight. The fix exists. But it kills usability. This isn't just OpenAI's problem. DeepMind and Tsinghua University reached identical conclusions working separately. Three elite AI labs. Independent research. Same result: this is permanent. Every time you get an answer from any LLM, you're not getting facts. You're getting the most statistically probable next words from a system that's been rewarded for never admitting when it's guessing. Is this real information, or just a confident hallucination? You can't know. And neither can the AI.

English
1
1
3
1.3K
Trong-Thang Pham
Trong-Thang Pham@trongthangpham·
If you're attending ICCV, I'd love to discuss our research on 3D volumetric scanpath modeling and its applications in medical imaging.
English
0
0
0
51
Trong-Thang Pham
Trong-Thang Pham@trongthangpham·
Excited to present our work at #ICCV2025! 🎉 I'll be presenting "CT-ScanGaze: A Dataset and Baselines for 3D Volumetric Scanpath Modeling" this Thursday morning (10:45 AM - 12:00 PM) at Exhibit Hall I, Poster #2003 (ID 181 in the main program). #ComputerVision #MedicalImaging
English
1
0
1
95
Trong-Thang Pham
Trong-Thang Pham@trongthangpham·
Summary: - First comprehensive dataset for gaze prediction on CT scan images with expert radiologist annotations - Novel deep learning approach combining spatial attention mechanisms with medical imaging expertise
English
1
0
0
43
Trong-Thang Pham
Trong-Thang Pham@trongthangpham·
Excited to share that our paper "CTScanGaze: A Comprehensive Dataset and Method for CT Scan Gaze Prediction" has been accepted as an ICCV 2025 Highlight Poster! Our paper got maximum scores (6,6,6) from all reviewers!!!
English
1
1
3
301
Trong-Thang Pham retweetledi
Jeremy Howard
Jeremy Howard@jeremyphoward·
The GPT 5 launch included a chart showing 52.8 as a bigger number than 69.1, which in turn is shown as the same magnitude as 30.8. Not quite ASI…
Jeremy Howard tweet media
English
56
81
892
95K
Trong-Thang Pham retweetledi
Tim Sweeney
Tim Sweeney@TimSweeneyEpic·
@kiaran_ritchie We have two simulation models. One that does exactly what you tell it, repeatable. And another that can do anything with infinite variation and creativity but unpredictably. The magic will happen when the two are brought together.
English
31
35
629
48.2K
Trong-Thang Pham retweetledi
Voxel51
Voxel51@Voxel51·
One of the biggest bottlenecks in deploying visual AI and computer vision is annotation, which can be both costly and time-consuming. Today, we’re introducing Verified Auto Labeling, a new approach to AI-assisted annotation that achieves up to 95% of human-level performance while cutting labeling costs by up to 100,000x and time by 5,000x. Read the full paper: arxiv.org/abs/2506.02359
English
3
200
114
12.1K
Trong-Thang Pham
Trong-Thang Pham@trongthangpham·
@DamienTeney @Michael_J_Black My trick is to let my body feel relax. I find that it is easy to rush things when I read in an uncomfortable place or position. Laying down in bed and proofreading is pretty effective for me.
English
0
0
0
7
Damien Teney
Damien Teney@DamienTeney·
@Michael_J_Black Also true for papers, but easier said than done🤷 I find it really difficult (for students & myself) to ignore what you already know (since you've written the whole thing!). The only trick I know is to leave the manuscript aside for >1 week before reading it through. Other tips?
English
6
1
34
3.5K
Michael Black
Michael Black@Michael_J_Black·
Writing your Ph.D. thesis? I have one request. Read every single word from beginning to end before giving it to your advisor. Every word. Straight through. You will immediately find things that are out of order, don't hang together, and don't form a coherent story.
English
13
18
413
45K
Trong-Thang Pham
Trong-Thang Pham@trongthangpham·
@sinenomine267 @karpathy I think your statement is a bit misleading here. Not all people who didn't like caffe is moving back to writing cuda. And the reason people didn't like Caffe is not only about writing C++.
English
0
0
0
38
Sine Nomine
Sine Nomine@sinenomine267·
@karpathy It is a bit funny that originally there were C++ frameworks like Caffe and people didn’t like how difficult it was to write C++ code - and moved to PyTorch and Tensorflow. Now the opposite is happening.
English
3
0
10
4.3K
Jia-Bin Huang
Jia-Bin Huang@jbhuang0604·
R2: While the results are impressive, this is a simple combination of diffusion transformer (ICCV 2023) and latent diffusion model (CVPR 2022). Limited novelty. Weak reject.
OpenAI@OpenAI

Introducing Sora, our text-to-video model. Sora can create videos of up to 60 seconds featuring highly detailed scenes, complex camera motion, and multiple characters with vibrant emotions. openai.com/sora Prompt: “Beautiful, snowy Tokyo city is bustling. The camera moves through the bustling city street, following several people enjoying the beautiful snowy weather and shopping at nearby stalls. Gorgeous sakura petals are flying through the wind along with snowflakes.”

English
52
144
1.6K
368.4K
Trong-Thang Pham retweetledi
Bo Wang
Bo Wang@BoWang87·
Yes, join us in the #CVPR2024 challenge: Segment Anything in Medical Images on Laptop! See the link: codabench.org/competitions/1… We welcome your contribution to pushing medical foundation models into real-world clinics!
Yuyin Zhou@yuyinzhou_cs

Wonderful work by @BoWang87 @JunMa_11 and the team! Join us in the #CVPR2024 Segment Anything in Medical Images on Laptop challenge (codabench.org/competitions/1…)! Looking forward to seeing innovations in building lightweight medical foundation models!

English
0
15
38
10.5K
Trong-Thang Pham retweetledi
Jon Barron
Jon Barron@jon_barron·
I've been applying ML to "classic" 3D/imaging vision problems for 15 years, so for 14 years I've been rebutting reviewers who don't like seeing _statistics_ (ew! scary!) applied to classic problems. Here's a snippet from a decade-old rebuttal that, regrettably, still holds up.
Jon Barron tweet media
English
10
25
375
46.7K