Junior_prompt_engineer

99.4K posts

Junior_prompt_engineer banner
Junior_prompt_engineer

Junior_prompt_engineer

@bert_on_spec

Hardhome Katılım Ağustos 2016
5.1K Takip Edilen1.5K Takipçiler
Junior_prompt_engineer retweetledi
Sumit
Sumit@_reachsumit·
Reasoning over Semantic IDs Enhances Generative Recommendation Proposes a two-stage framework that enables LLMs to reason over discrete item tokens for generative recommendation, using enriched SID-language alignment and RL. 📝 arxiv.org/abs/2603.23183 👨🏽‍💻 github.com/HappyPointer/S…
English
0
5
13
570
Junior_prompt_engineer retweetledi
Sumit
Sumit@_reachsumit·
OneSearch-V2: The Latent Reasoning Enhanced Self-distillation Generative Search Framework Kuaishou presents a generative search framework that enhances complex query understanding. 📝 arxiv.org/abs/2603.24422 👨🏽‍💻 github.com/benchen4395/on…
English
0
3
15
645
Junior_prompt_engineer retweetledi
Zhuofeng Li
Zhuofeng Li@zhuofengli96475·
🚀 OpenResearcher paper is finally released! 🔥 We explore how to synthesize long-horizon research trajectories for deep-research agents — fully offline, scalable, and low-cost, without relying on live web APIs. 📄 huggingface.co/papers/2603.20… 🧩Two key ideas: Offline Corpus — One-time bootstrapping seeds 10K gold passages + 15M-doc FineWeb corpus. 📚 Explicit Browsing Primitives — Just 3 ops: search / open / find. The agent learns not just what to retrieve, but how to inspect docs and localize evidence at multiple scales. 🔎 📊 Results: 54.8% on BrowseComp-Plus with our 30B-A3B — #1 open-source under the same search engine setup. Beating much larger models like GPT-4.1, Claude-Opus-4, Gemini-2.5-Pro, and DeepSeek-R1. 💡 Insights: Beyond accuracy, we dissect deep research pipeline design—from data filtering and agent configuration to retrieval accuracy dynamics (RQ1-RQ5). Try it yourself: 🛠️ Code: github.com/TIGER-AI-Lab/O… 🤗 Models & data: huggingface.co/collections/TI… 🚀 Demo: huggingface.co/spaces/OpenRes… #llms #agentic #deepresearch #tooluse #opensource #retrieval #SFT
Zhuofeng Li tweet mediaZhuofeng Li tweet mediaZhuofeng Li tweet mediaZhuofeng Li tweet media
Dongfu Jiang@DongfuJiang

🚀 Introducing OpenResearcher: a fully offline pipeline for synthesizing 100+ turn deep-research trajectories—no search/scrape APIs, no rate limits, no nondeterminism. 💡 We use GPT-OSS-120B + a local retriever + a 10T-token corpus to generate long-horizon tool-use traces (search → open → find) that look like real browsing, but are free + reproducible. 📈 The payoff: SFT on these trajectories turns Nemotron-3-Nano-30B-A3B from 20.8% → 54.8% accuracy on BrowseComp-Plus (+34.0). 🧩 What makes it work? 🔎 Offline corpus = 15M FineWeb docs + 10K “gold” passages (bootstrapped once) 🧰 Explicit browsing primitives = better evidence-finding than “retrieve-and-read” 🎯 Reject sampling = keep only successful long-horizon traces 🧵 And we’re releasing everything: ✅ code + search engine + corpus recipe ✅ 96K-ish trajectories + eval logs ✅ trained models + live demo 👨‍💻 GitHub: github.com/TIGER-AI-Lab/O… 🤗 Models & data: huggingface.co/collections/TI… 🚀 Demo: huggingface.co/spaces/OpenRes… 🔎 Eval logs: huggingface.co/datasets/OpenR… #llms #agentic #deepresearch #tooluse #opensource #retrieval #SFT

English
11
60
306
42.8K
Junior_prompt_engineer retweetledi
Sumit
Sumit@_reachsumit·
KARMA: Knowledge-Action Regularized Multimodal Alignment for Personalized Search at Taobao Alibaba identifies Semantic Collapse in LLM-based personalized search and proposes train-only decodability regularization to preserve semantic knowledge. 📝 arxiv.org/abs/2603.22779
English
1
2
12
366
Junior_prompt_engineer retweetledi
Vaishnavi
Vaishnavi@_vmlops·
awesome-mcp-servers is the only MCP resource list you need covers everything: → browser automation → cloud platforms (AWS, K8s, Cloudflare) → databases, dev tools, file systems → AI agents, search, monitoring & more github.com/punkpeye/aweso…
Vaishnavi tweet media
English
1
9
56
2.6K
Junior_prompt_engineer retweetledi
Chaumian
Chaumian@chaumian·
Zero-Shot Vulnerability Detection in Low-Resource Smart Contracts Through Solidity-Only Training arxiv.org/abs/2603.21058
English
1
3
9
347
Junior_prompt_engineer retweetledi
fly51fly
fly51fly@fly51fly·
[CL] Measuring Reasoning Trace Legibility: Can Those Who Understand Teach? D Roytburg, S Sridhar, D Ippolito [CMU] (2026) arxiv.org/abs/2603.20508
fly51fly tweet mediafly51fly tweet mediafly51fly tweet mediafly51fly tweet media
English
0
4
16
957
Junior_prompt_engineer retweetledi
Sumit
Sumit@_reachsumit·
A Brief Comparison of Training-Free Multi-Vector Sequence Compression Methods @Robro612 et al. evaluate training-free token pruning vs. pooling for multi-vector retrieval, finding token merging strictly superior for reducing index size. 📝 arxiv.org/abs/2603.22434
English
0
7
34
2.7K
Junior_prompt_engineer retweetledi
Arjun
Arjun@arjunkocher·
Exclusive Self Attention (XSA). paper breakdown: k-a.in/XSA.html
Arjun tweet media
English
5
34
429
18.8K
Junior_prompt_engineer retweetledi
Sumit
Sumit@_reachsumit·
SkillRouter: Retrieve-and-Rerank Skill Selection for LLM Agents at Scale Alibaba proposes a 1.2B retrieve-and-rerank pipeline for selecting skills from ~80K pools, showing that full skill body text is the decisive signal for selection. 📝 arxiv.org/abs/2603.22455
English
4
13
139
7.9K
Junior_prompt_engineer retweetledi
Sumit
Sumit@_reachsumit·
GEM: A Native Graph-based Index for Multi-Vector Retrieval Presents a native graph-based indexing framework for multi-vector retrieval that constructs a proximity graph directly over vector sets and achieves up to 16x speedup. 📝arxiv.org/abs/2603.20336 👨🏽‍💻github.com/sigmod26gem/si…
English
0
6
31
1.1K
Junior_prompt_engineer retweetledi
fly51fly
fly51fly@fly51fly·
[AI] Demonstrations, CoT, and Prompting: A Theoretical Analysis of ICL X Tong, Y Zeng, J Zhang [Microsoft Research & University of Wisconsin-Madison] (2026) arxiv.org/abs/2603.19611
fly51fly tweet mediafly51fly tweet mediafly51fly tweet media
English
0
6
16
963
Junior_prompt_engineer retweetledi
Xingyi Yang
Xingyi Yang@yxy2168·
Interesting paper: Exclusive Self Attention (XSA). The key idea is simple: force attention to model information orthogonal to self-value. A nice example of how small changes to the attention mechanism can still matter. arxiv.org/abs/2603.09078
Xingyi Yang tweet media
English
0
26
161
9.2K
Junior_prompt_engineer retweetledi
Sudo su
Sudo su@sudoingX·
let me get you started in local AI and bring you to the edge. if you have a GPU or thinking about diving into the local LLM rabbit hole, first thing you do before any setup is join x/LocalLLaMA. this is the community that will help you at every step. post your issue and we will direct you, debug with you, and save you hours of work. once you're in, follow these three: @TheAhmadOsman the oracle. this is where you consume the latest edges in infrastructure and AI. if something dropped you hear it from him first. his content alone will keep you ahead of most. @0xsero one man army when it comes to model compression, novel quantization research, new tools and tricks that make your local setup better. you will learn, experiment, and discover things you didn't know existed. @Teknium maker of Hermes Agent, the agent i use every day from @NousResearch. from Teknium you don't just stay at the frontier, you get your hands on the tools before everyone else. this is where things are headed. if you follow me follow these three and join the community. you will be ahead of most people in this space. if you run into wrong configs, stuck debugging hardware, or can't get a model to load, post there so we can help. get started with local AI now. not only understand the stack but own your cognition. don't pay openai fees on top of giving them your prompts, your research, and your most valuable thinking to be monitored and metered. buy a GPU and build your own token factory.
Sudo su tweet media
English
60
60
800
96.6K
Junior_prompt_engineer retweetledi
Sumit
Sumit@_reachsumit·
A Super Fast K-means for Indexing Vector Embeddings @LeonardoKuffo et al. introduce a k-means variant that prunes unnecessary dimensions during clustering, achieving faster indexing than FAISS on CPUs and cuVS on GPUs. 📝 arxiv.org/abs/2603.20009 👨🏽‍💻 github.com/cwida/SuperKMe…
English
0
7
42
1.6K