0x_arjunghosh_ai

79.5K posts

0x_arjunghosh_ai banner
0x_arjunghosh_ai

0x_arjunghosh_ai

@arjunghosh

Non-Linear #Thinker, #Futurist, #solopreneur, Crafting #AI & multi #AgenticAI, Coach, Speaker, Founder at https://t.co/2FxweBxUFG, Chief AI & Tech Officer @flexilytics

Mumbai & Kolkata Katılım Mayıs 2007
5.3K Takip Edilen3.3K Takipçiler
0x_arjunghosh_ai retweetledi
Mayank Pratap Singh
Mayank Pratap Singh@Mayank_022·
𝐕𝐢𝐬𝐮𝐚𝐥 𝐛𝐥𝐨𝐠 on Vision Transformers is live. vizuaranewsletter.com/p/vision-trans… Learn how ViT works from the ground up, and fine-tune one on a real classification dataset. CNNs process images through small sliding filters. Each filter only sees a tiny local region, and the model has to stack many layers before distant parts of an image can even talk to each other. Vision Transformers threw that whole approach out. ViT chops an image into patches, treats each patch like a token, and runs self-attention across the full sequence. Every patch can attend to every other patch from the very first layer. No stacking required. That global view from layer one is what made ViT surpass CNNs on large-scale benchmarks. 𝐖𝐡𝐚𝐭 𝐭𝐡𝐞 𝐛𝐥𝐨𝐠 𝐜𝐨𝐯𝐞𝐫𝐬: - Introduction to Vision Transformers and comparison with CNNs - Adapting transformers to images: patch embeddings and flattening - Positional encodings in Vision Transformers - Encoder-only structure for classification - Benefits and drawbacks of ViT - Real-world applications of Vision Transformers - Hands-on: fine-tuning ViT for image classification The Image below shows Self-attention connects every pixel to every other pixel at once. Convolution only sees a small local window. That's why ViT captures things CNNs miss, like the optical illusion painting where distant patches form a hidden face. The architecture is simple. Split image into patches, flatten them into embeddings (like words in a sentence), run them through a Transformer encoder, and the class token collects info from all patches for the final prediction. Patch in, class out. Inside attention: each patch (query) compares itself to all other patches (keys), softmax gives attention weights, and the weighted sum of values produces a new representation aware of the full image, visualizes what the CLS token actually attends to through attention heatmaps. The second half of the blog is hands-on code. I fine-tuned ViT-Base from google (86M params) on the Oxford-IIIT Pet dataset, 37 breeds, ~7,400 images. 𝐁𝐥𝐨𝐠 𝐋𝐢𝐧𝐤 vizuaranewsletter.com/p/vision-trans… 𝐒𝐨𝐦𝐞 𝐑𝐞𝐬𝐨𝐮𝐫𝐜𝐞𝐬 Dr @sreedathpanat Videos on ViT ViT paper dissection youtube.com/watch?v=U_sdod… Build ViT from Scratch youtube.com/watch?v=ZRo74x… Original Paper arxiv.org/abs/2010.11929 Next up: demystifying Low-Rank Adaptation (LoRA) in PEFT! Follow me @Mayank_022 along for more deep learning insights, cool fine-tuning projects, and updates from the upcoming blog posts.
YouTube video
YouTube
YouTube video
YouTube
GIF
English
13
346
2.2K
83.4K
Harj Taggar
Harj Taggar@harjtaggar·
Does anyone know what’s going on with the lobster on Wall Street lol?
Harj Taggar tweet media
English
262
149
1.8K
1.5M
0x_arjunghosh_ai retweetledi
Guilhem Herail
Guilhem Herail@guilhemherail·
At @ycombinator Alumni Demo Day! W26 is an unfair concentration of talent 🤯
Guilhem Herail tweet media
English
1
5
31
8.3K
0x_arjunghosh_ai retweetledi
Charly Wargnier
Charly Wargnier@DataChaz·
Wow. @GarryTan (@ycombinator's CEO) just dropped the ultimate cheat code for software engineers. 🔥 He just open-sourced gstack, his personal toolkit that transforms Claude Code from a basic chatbot into an entire virtual engineering department. Instead of asking Claude to "build a feature" and hoping for the best, gstack lets you summon specific "brains" on demand: → The Visionary: /plan-ceo-review acts like Brian Chesky. It stops you from building boring features and pushes for magic. → The Architect: /plan-eng-review draws sequence diagrams and state machines before coding. → The Paranoid Reviewer: /review looks for N+1 queries and stale reads. → The QA Lead: /qa literally logs into your staging environment, clicks around, takes screenshots, and gives your app a health score in 60 seconds. The QA tool alone is built on Bun and Playwright, running 20x faster than Claude’s native Chrome MCP with zero context bloat. He used this exact setup to ship 100 PRs a week for the last 50 days. Get the repo in 🧵 ↓
Charly Wargnier tweet media
English
20
51
352
67.4K
0x_arjunghosh_ai retweetledi
Bloomberg Technology
Bloomberg Technology@technology·
"I realized tech is this thing that can bring people out of whatever situation they're in and often into prosperity. And that's what I want for everyone." @ycombinator’s @garrytan tells @emilychangtv how tech changed his family's life. Watch here: trib.al/sxg1VGR
English
51
138
829
3.8M
Kevin Rose
Kevin Rose@kevinrose·
and the ref=gstack link to track incoming from people actually coding... so well executed.
English
3
1
18
8.9K
0x_arjunghosh_ai retweetledi
Kevin Rose
Kevin Rose@kevinrose·
possibly the sharpest VC marketing move I've seen... @garrytan ships 15 claude code skills, the repo hits 37k stars and 4.6k forks, then -- only after delivering real value - drops the pitch, bravo 👏:
Kevin Rose tweet media
English
48
37
779
114K
Tom Blomfield
Tom Blomfield@t_blom·
Pretty soon, I think we’ll see software shipping with Claude Code SDK embedded inside. Users will use it to configure and modify the software to meet their exact needs. The best changes will get passed back to the software developer and reincorporated in the master release.
English
228
52
1.2K
387.7K
0x_arjunghosh_ai retweetledi
Garry Tan
Garry Tan@garrytan·
Weird realization: The best AI coding is in the morning when you are fresh from a night full of dreaming about latent space. Sleep early. Wake up early. The best ideas are in the morning. It's not just about raw token maxxing. It is about teaching the machines the right abstraction that comes out of your own personal experience and the synthesis that comes from a good night's sleep.
English
223
69
1.4K
80.7K
0x_arjunghosh_ai
0x_arjunghosh_ai@arjunghosh·
WTF @facebook, why can't u still fix the simple asynchronous JS loader "see more" for ur Birthday wishes page on FB wall ur app?After 4-4 batches it started crashing & Why not even an infinite scroll? In this day & age of #vibe coding, should I send u guys the code @MetaforDevs ?
0x_arjunghosh_ai tweet media
English
0
0
0
25
0x_arjunghosh_ai
0x_arjunghosh_ai@arjunghosh·
@CodeByPoonam It was always so and so was any android phone app! Text scam mode was always there 🤦😎
English
0
0
0
257
Poonam Soni
Poonam Soni@CodeByPoonam·
Google Drive just made every document scanner app on your phone irrelevant.
English
11
15
116
12.8K
0x_arjunghosh_ai
0x_arjunghosh_ai@arjunghosh·
Ofocuse why would LLM model not?Let me ask u as Human Intelligence to talk,read & solve a math problems written 4 you in Nagamese language,u will also brainfreeze & zero shot #epicfail 😎 I mean do u really not listen to @ylecun & his rant that LLMs r just hyper text predictors?
Lossfunk@lossfunk

🚨 Shocking: Frontier LLMs score 85-95% on standard coding benchmarks. We gave them equivalent problems in languages they couldn't have memorized. They collapsed to 0-11%. Presenting EsoLang-Bench. Accepted to the Logical Reasoning and ICBINB workshops at ICLR 2026 🧵

English
0
0
0
21
0x_arjunghosh_ai
0x_arjunghosh_ai@arjunghosh·
Ofocuse why would LLM model not?Let me ask u as Human Intelligence to talk,read & solve a math problems written 4 you in Nagamese language,u will also brainfreeze & zero shot #epicfail 😎 I mean do u really not listen to @ylecun & his rant that LLMs r just hyper text predictors?
Paras Chopra@paraschopra

We found a task where LLMs struggle massively! Give them a coding problem in Python and they'd work great. Give the same problem in brainfuck and zero-shot their performance is ~0% +[--------->+<]>+.++[--->++<]>+.

English
0
0
0
22
0x_arjunghosh_ai
0x_arjunghosh_ai@arjunghosh·
@paraschopra Ofocuse why would LLM model not?Let me ask u as Human Intelligence to talk,read & solve a math problems written 4 you in Nagamese language,u will also brainfreeze & zero shot #epicfail 😎 I mean do u really not listen to @ylecun & his rant that LLMs r just hyper text predictors?
English
0
0
0
13
Paras Chopra
Paras Chopra@paraschopra·
We found a task where LLMs struggle massively! Give them a coding problem in Python and they'd work great. Give the same problem in brainfuck and zero-shot their performance is ~0% +[--------->+<]>+.++[--->++<]>+.
Lossfunk@lossfunk

🚨 Shocking: Frontier LLMs score 85-95% on standard coding benchmarks. We gave them equivalent problems in languages they couldn't have memorized. They collapsed to 0-11%. Presenting EsoLang-Bench. Accepted to the Logical Reasoning and ICBINB workshops at ICLR 2026 🧵

English
90
33
1K
175.4K