Someshwaran Mohankumar

154 posts

Someshwaran Mohankumar banner
Someshwaran Mohankumar

Someshwaran Mohankumar

@som23x

Developer Advocate @elastic! Open-source enthusiast, love to collaborate and share knowledge. I enjoy coding, troubleshooting, and blogging Tech!

Chennai, India Katılım Ocak 2012
267 Takip Edilen63 Takipçiler
Someshwaran Mohankumar
@saen_dev @mdancho84 Nowadays, for NER, the transformers are doing a much better job at scale. And, in terms of PDF, it became a default training set. Can expect better performance. Will try this LangExtract and will share some results.
English
0
0
0
24
Saeed Anwar
Saeed Anwar@saen_dev·
@mdancho84 Open source document extraction that actually works would kill a lot of enterprise pricing. The real test is how it handles scanned PDFs with mixed layouts because that's where every extraction tool falls apart in production.
English
1
0
1
1.1K
Matt Dancho (Business Science)
RIP document extractors. Google just released LangExtract: Open-source. Free. Better than $100K enterprise tools. Here’s what it does: 🧵
Matt Dancho (Business Science) tweet media
English
18
113
863
72.3K
Someshwaran Mohankumar
@TheSuperEng @drummatick When DeepSeekv3 launched, the West strategically moved to restrict NVDA chip exports from the US as it became very powerful. If similar progress is now happening through Chinese alternatives for Mythos, just imagine the impact 😂
English
1
0
2
11
Shubh Srivastava
Shubh Srivastava@TheSuperEng·
@drummatick This is what I said earlier lol. Wait until they open source mythos level AI.
English
1
0
1
29
Saurabh Kumar
Saurabh Kumar@drummatick·
"DeepSeek is a threat." "It's a security concern." "It's Chinese." Dude, go ahead and ask American companies to do more open source. We have cheaper models, free access to open weights, and don't need to rely fully on monopolies, thanks to Open weight models.
English
2
0
25
643
Someshwaran Mohankumar
I think part of the reason we got into this situation is the tooling. The AI ecosystems like LangChain, LangGraph, LlamaIndex, and AutoGen (and more) made it easy to build agentic RAG systems. So we as devs, pushed to a thought process, what if, if the pipeline we build today need flexibility and scale in the future, even when a simple RAG + function call would’ve been enough. Feels like using a sledgehammer to crack a nut.
English
0
0
0
15
Someshwaran Mohankumar
Was discussing this with a friend building RAG pipelines. Not every problem needs Agentic RAG, even simple function calls are enough. Feels like the K8s-on-a-resume situation: powerful, but it's an overkill for an "Hello World" application. No need to burn tokens when you know the answer is right in front of you! 😄
English
1
0
2
173
Leonie
Leonie@helloiamleonie·
This is the most common question I’ve been asked recently: “Should I replace my RAG system with an agentic RAG one?” Well, are you happy with your current RAG system’s performance? Then, probably not. The key difference between traditional RAG and agentic RAG is HOW the context is built: Traditional RAG has a fixed retrieval pipeline that retrieves exactly once. → The LLM is a passive recipient of additional context Agentic RAG has access to one or more retrieval tool(s) to retrieve on demand. → The agent actively builds its own context > Does your use case always require exactly one retrieval step? (Example: always pull in customer information) > Or do you have cases where no additional context is needed? > Or do you have cases where multi-hop retrieval is necessary? If you don’t know, set up a simple evaluation with a gold test set of a few common user patterns: > Do precision and recall improve for your retrieval component? > How much does the end-to-end performance improve based on performance improvements of the retrieval component? > Do these performance improvements justify the added latency and cost of agentic RAG?
Leonie tweet media
English
4
8
33
1.8K
Someshwaran Mohankumar
@drummatick I've been falling back and forth with this thought process off-late, to use SKILLS.md or just using the function or tool call directly that saves my time. I completely agree with you. Given there's no abstraction needed for all minimal efforts or routine tasks.
English
0
0
0
49
Saurabh Kumar
Saurabh Kumar@drummatick·
My prediction is, slowly everyone will realize that using SKILLS to implement a predictable workflow like browsing, checking logs, or testing use just wasting useless tokens and just having the raw code is better than have SKILLS.md later converted into the code
English
7
0
30
920
Someshwaran Mohankumar
@vavanessadev @Varnika99 @parthfullstack If you have more control over the memory layer, you can index those episodic memories to a persistent layer like an index, create a connector/tool/MCP to sync between Claude and Codex. This got more success when you switch. Kinda of vendor neutral 😅
English
1
0
2
25
Varnika
Varnika@Varnika99·
Why is everyone shifting from claude codex? Should I also consider it?
English
104
1
110
8.4K
Someshwaran Mohankumar
Exploring a local-first agent setup with continue.dev + Ollama + Gemma4:26b (MoE) with M4 Pro, 48GB. Any suggestions before I go deeper?
Someshwaran Mohankumar tweet media
English
1
0
3
140
Someshwaran Mohankumar
@aravindputrevu Yeh toh irony hai! 😂 Technology kitni bhi aage jaaye, containers ho, cloud ho, ya ab LLMs, andar se wahi AWK, SED, aur Bash chal raha hai. Linux fundamentals ek baar seekh lo, zindagi bhar kaam aate hain. Low-level kabhi nahi badalta, sirf uske upar ka layer badalta hai.
हिन्दी
0
0
1
20
Aravind Putrevu
Aravind Putrevu@aravindputrevu·
Our Unix professor in college made us write AWK scripts as part of coursework. We felt that who needs this today, lol? We thought we would end up writing in a new-age programming language like Python. Fast-forward to 2026 - staring at LLMs writing scripts in SED and AWK 🙂 > echo PRESENT
English
1
0
11
466
Zain Shah
Zain Shah@zan2434·
@code_rgb @eddiejiao_obj @drewocarr will put out a more in depth explanation on how it works in a few days! but long story short a lot of activation caching, quantization, and torch.compile + memory snapshotting
English
3
2
15
4.5K
Zain Shah
Zain Shah@zan2434·
Imagine every pixel on your screen, streamed live directly from a model. No HTML, no layout engine, no code. Just exactly what you want to see. @eddiejiao_obj, @drewocarr and I built a prototype to see how this could actually work, and set out to make it real. We're calling it Flipbook. (1/5)
English
1.1K
3.4K
26.5K
5.6M
Someshwaran Mohankumar
@zan2434 @eddiejiao_obj @drewocarr Does a new query always reset the state? Trying to figure out if we can persist state within the same session or if it’s all just temporary caching. If I clear a query and run it again, does the browser cache pick it up? To avoid gen. twice the same query!
English
0
0
0
8
Someshwaran Mohankumar
@zan2434 @eddiejiao_obj @drewocarr This is a wonderful representation how Text > Image > Video > Interactive pixels. I see when I enter a query, it creates draft version as an Image and then the model goes into turning it into a video.
English
1
0
0
8
Naveen
Naveen@the_mcnaveen·
> builds android app > hit $100 > now build ios app > millionaire soon
Naveen tweet media
English
5
0
8
187
Philipp Krenn
Philipp Krenn@xeraa·
took some time off in preparation for las vegas; and #GoogleCloud — busy week ahead but looking forward to seeing lots of people
Philipp Krenn tweet mediaPhilipp Krenn tweet mediaPhilipp Krenn tweet mediaPhilipp Krenn tweet media
English
1
0
13
281
Someshwaran Mohankumar retweetledi
Elastic Dev
Elastic Dev@elastic_devs·
Most agent knowledge bases go stale faster than people realize. If you’re periodically re-embedding docs, your retrieval layer is already drifting out of date. Elastic Workflows fix this with a live context pipeline: Trigger → crawl (Tavily, Firecrawl) → transform → index → retrieve at runtime Your agents always query fresh context instead of snapshots.
Elastic Dev tweet media
English
0
1
9
388