Adetayo

2.2K posts

Adetayo banner
Adetayo

Adetayo

@tee_sharp_

Mobile Software Engr. | Jesus Saves | God bless Nigeria🇳🇬 | Rust | Substrate

Secret Place Katılım Nisan 2013
1.3K Takip Edilen421 Takipçiler
Sabitlenmiş Tweet
Adetayo
Adetayo@tee_sharp_·
Never again.
English
1
0
1
0
Adetayo retweetledi
Adetayo retweetledi
Dr. Harri
Dr. Harri@Harri_obi·
I hope the Hyperbridge team bounces back from this.
English
27
6
169
16.5K
Adetayo retweetledi
Adetayo retweetledi
andy
andy@b1rdmania·
15 more London tech startups doing interesting shit: • @facultyai - AI deployment for enterprises • @ManticGames - AI sports prediction • @tldraw - infinite canvas for thinking • @surrealdb - multi-model database in Rust • @lightdash_devs - open-source BI for dbt • @jackandjillai - AI conversation design 4 recruiting • @encodeclub - web3 education • @zep_ai - long-term memory for AI • @dust4ai - AI assistants for teams • @spice_ai - time series AI infrastructure • @Replit - AI-powered coding (UK-founded) • @graphcore - AI chips • @Papercup_AI - AI dubbing for video • @humanloop - prompt engineering & LLM ops • @AUAR_official - robots building buildings (again, I like it so much) starting to feel like we should throw together some kind of conference... and if you're building something interesting in London drop it below - adding to the map next week and building on api / data layers so it's a bit more useful. londonmaxxxing.com
English
17
11
162
9.4K
Adetayo retweetledi
ℏεsam
ℏεsam@Hesamation·
Harvard's just open-sourced their ML Systems textbook. it's extremely practical for not just learning how to build and train models, but to build production systems (the skill that actually matters). topics are cool af: > building autograd, optimizers, attention, and a mini-pytorch from scratch to truly learn how an ML framework runs. (i love this the most) > basics of DL, batch sizes, precision, model architectures, and training > ML performance optimization, HW acceleration, benchmarking, efficiency so this is not just an intro to machine learning, it's the full package from the beginning to the actual end. right now you can read the book and access the code for free. this is one of the best books I've seen dropping in 2025, so don't sleep on it. here's the repo (you can find the book link there): github.com/harvard-edge/c…
ℏεsam tweet media
English
50
659
5.2K
279.5K
Adetayo retweetledi
Mihail Eric
Mihail Eric@mihail_eric·
All assignments for Stanford's The Modern Software Developer are now available online. This is the first comprehensive university course covering how coding LLMs are transforming every stage of the software development life cycle. The assignments are intended to take you from noob to expert in how to use AI to improve your software engineering productivity. Enjoy! github.com/mihail911/mode…
English
28
254
1.9K
273.2K
Adetayo retweetledi
anshuman
anshuman@athleticKoder·
Techniques I'd master to build great evals for AI apps. 1. LLM-as-Judge 2. Reference-based similarity metrics 3. Pairwise comparison tournaments 4. Human-in-the-loop evaluation 5. Synthetic data generation 6. Adversarial test case creation 7. Multi-dimensional rubrics 8. Regression testing on golden datasets 9. A/B testing with live traffic 10. Statistical significance testing 11. Evaluation dataset curation & versioning 12. Domain-specific benchmarks 13. Red teaming & jailbreak testing 14. Latency & cost monitoring 15. User feedback loops 16. Calibration & confidence scoring
English
10
41
307
15.8K
Miko
Miko@Mho_23·
i built a tool that allows me to clone the brain of any youtuber… i used it on alex hormozi so i can ask him business questions at 3am and get answers word-for-word like he'd give them works with anyone: gary vee, naval, graham stephan, any creator in your niche i put together: - the tool that uploads entire channels automatically - my 5-minute setup process - prompt framework for answers that match their exact style RT + reply 'MENTOR' and i'll send you everything (must follow so i can dm)
English
2.2K
1.1K
3.9K
814.4K
Adetayo retweetledi
Hyperbridge
Hyperbridge@hyperbridge·
Web3 autonomy starts here. Sovereign Intents coming soon.
English
72
1.4K
769
167.7K
Adetayo retweetledi
anshuman
anshuman@athleticKoder·
You’re in a Machine Learning interview at OpenAI, and the interviewer asks: “Why is everyone switching from RLHF to DPO? Isn’t RLHF the proven approach?” Here’s how you answer: Don’t say: “DPO is simpler” or “RLHF is too complex.” Too surface-level. The real answer is the reward model bottleneck. RLHF trains a separate reward model that becomes a noisy proxy for human preferences. DPO directly optimizes the policy on preference data. You’re eliminating the broken telephone. Here’s why RLHF is fundamentally flawed: Your training pipeline: Human preferences → Train reward model → Use RL to optimize policy against reward model. Problem: The reward model is trained on limited data (10k-100k comparisons), but then used to generate 1M+ training signals. It’s overconfident on out-of-distribution outputs. Reward model accuracy ≠ Alignment quality. btw subscribe to my newsletter to get these posts for free - fullstackagents.substack.com The RLHF failure modes are brutal: > Reward hacking: Model finds adversarial outputs that score high but are gibberish > Mode collapse: Policy degenerates to only generate “safe” high-reward outputs > Reward model brittleness: 75% accuracy on test set → 100% confident predictions in RL > Training instability: PPO hyperparameters require black magic to converge You’re building a skyscraper on quicksand. One unstable component breaks everything. The complexity comparison: RLHF pipeline: - SFT on demonstrations (1 week) - Train reward model on preferences (2 days) - PPO training against reward model (1-2 weeks, often fails) - Extensive hyperparameter tuning (pray to the RL gods) DPO pipeline: - SFT on demonstrations (1 week) - Train directly on preferences (2 days) Done. No RL, no reward model. RLHF: 3+ weeks, unstable. DPO: 9 days, stable. The fundamental difference that matters: RLHF objective: > maximize E[reward_model(policy(x))] - β × KL(policy || base) > Requires RL (PPO/REINFORCE) > Reward model is separate neural net > Unstable optimization landscape DPO objective: > maximize log(σ(β × log(π(y_w|x)/π(y_l|x)/π_ref))) > Direct supervised learning on preferences > No reward model needed > Stable gradient descent That eliminated reward model changes everything. No more broken telephone. The performance gap that surprised everyone: MT-Bench scores (GPT-4 as judge): Llama 2 base: 4.2/10 Llama 2 + RLHF: 6.9/10 Llama 2 + DPO: 7.1/10 DPO beats RLHF despite being “just” supervised learning. The simplicity IS the feature.
English
11
26
322
33K
Adetayo retweetledi
Kuba ✨
Kuba ✨@kubadesign·
footer designs
Kuba ✨ tweet mediaKuba ✨ tweet mediaKuba ✨ tweet mediaKuba ✨ tweet media
English
52
37
920
74.3K
Adetayo retweetledi
Akshay 🚀
Akshay 🚀@akshay_pachaar·
How LLMs work under the hood? This is the best place to visually understand the internal workings of a transformer-based LLM. Explore tokenization, self-attention, and more in an interactive way:
English
9
168
870
66.1K
Adetayo retweetledi
Himanshu
Himanshu@himanshuq14·
Use this structure for your SAAS Landing page Thank me later
Himanshu tweet media
English
152
610
9.3K
732.3K
Adetayo retweetledi
Akshay 🚀
Akshay 🚀@akshay_pachaar·
Everyone is sleeping on this new OCR model! dots-ocr is a new 1.7B vision-language model that achieves SOTA performance on multilingual document parsing. - Supports 100+ languages - Works with both images and PDFs - Handles text, tables, formulas seamlessly 100% open-source.
English
42
456
3.1K
251.8K
Adetayo retweetledi
Akshay 🚀
Akshay 🚀@akshay_pachaar·
Google just dropped a new LLM! You can run it locally on just 0.5 GB RAM. Let's fine-tune this on our own data (100% locally):
English
184
1.2K
14.4K
2M
Adetayo retweetledi
Alex Prompter
Alex Prompter@alex_prompter·
what are large language models actually doing? i read the 2025 textbook "Foundations of Large Language Models" by tong xiao and jingbo zhu and for the first time, i truly understood how they work. here’s everything you need to know about llms in 3 minutes↓
Alex Prompter tweet media
English
81
907
7K
1.1M
Adetayo retweetledi
Web3 Philosopher
Web3 Philosopher@seunlanlege·
Real ones know that the first million is the hardest. Higher.
Web3 Philosopher tweet media
English
31
30
255
14.3K
Adetayo retweetledi
Tobi Akinpelu
Tobi Akinpelu@tobiakinpelu_·
Since 2023, We’ve had 500,000+ newsletter views on engaged 8,000+ subscribers. We’ve spotlighted 100+ entrepreneurs and creators — without charging a penny. We do this for networking. Now opening up to support the next 10,000. Like, RT, Get featured: forms.gle/LBqXSA9d3NHvj5…
Tobi Akinpelu tweet mediaTobi Akinpelu tweet mediaTobi Akinpelu tweet media
English
0
5
9
944