Boochi 🇳🇬

824 posts

Boochi 🇳🇬 banner
Boochi 🇳🇬

Boochi 🇳🇬

@boochi_dot_dev

• ICU Nurse • Computer Scientist • NeuralMind • ML Engineer

Portsmouth, England Katılım Ekim 2017
602 Takip Edilen300 Takipçiler
Boochi 🇳🇬 retweetledi
mei
mei@multiply_matrix·
at a bare minimum, the scientists building opensource AI deserve their credit where it is due. scientists devote entire careers to advancing the frontier of knowledge. academic ethics exist for a reason. in school, if you forget to cite your sources, it's called plagiarism.
English
0
2
12
576
Boochi 🇳🇬
Boochi 🇳🇬@boochi_dot_dev·
@FeiHao88 This is getting interesting… So, after observing that the perplexity of a given model closely matches the distribution of a company’s user base (as in Cursor’s), which metric would you use to select a model for CPT?
English
0
0
0
7
Google AI Studio
Google AI Studio@GoogleAIStudio·
What are you vibe coding this weekend?
English
539
53
1.4K
146.4K
Boochi 🇳🇬
Boochi 🇳🇬@boochi_dot_dev·
What’s a better way to evaluate a base model? If you have an ensemble of pre-trained LLMs/LLM checkpoints, your perplexity score is the most reliable metric to determine which has the strongest performance on the language you wish to further improve its ability via post-training. Cursor is a software coding, and using perplexity rating across a set of programming language tasks seems like the most stable option
English
1
0
0
860
Boochi 🇳🇬
Boochi 🇳🇬@boochi_dot_dev·
I hate to admit it, but functional programmers were right all along. OOP sucks, man!
English
0
0
0
11
Tofunmi🌸
Tofunmi🌸@Tofunmithedev·
No syntax is beating that of Python!
English
67
37
252
10.9K
Boochi 🇳🇬
Boochi 🇳🇬@boochi_dot_dev·
@SooYes @kevinsxu Lol, I don’t know how you rank world class teams but ByteDance and Baidu have way more competent research teams than DeepSeek and Qwen💯
English
1
0
0
65
SYS
SYS@SooYes·
@kevinsxu Qwen team is probably the only world class LLM research team in China other than Deepseek. Base on what I heard, he probably made the decision due to benchmarks. Specifically that qwen3.5 lagging behind on benchmark. But the mistake is huge.... Leading to entire team leaving.
English
3
0
9
1.8K
Kevin S. Xu
Kevin S. Xu@kevinsxu·
Alibaba said nothing about open source as part of its future AI strategy in its earnings call I thought it would at least pay some temporary lip service to open source Qwen, as we know it, is dead
English
9
22
256
140.1K
omo
omo@sebiomo_·
I just automated video content generation for TikTok. 😙 Supabase → OpenAI Script → OpenAI TTS → Pexels Video → Remotion Render → MP4 One command. ~$0.03 per video.
English
44
31
550
93.9K
Boochi 🇳🇬 retweetledi
Mustafa
Mustafa@oprydai·
the jacobian matrix is how multivariable systems actually move you don’t deal with one variable anymore you deal with transformations input vector → output vector the jacobian captures how every input dimension affects every output dimension what it is: → a matrix of partial derivatives → each row = one output function → each column = one input variable J(i,j) = ∂f_i / ∂x_j why it matters: → it’s the local linear approximation of a nonlinear system → it tells you how small changes propagate → it converts messy systems into something you can compute in physics: → coordinate transformations → velocity mappings → change of variables in integrals in robotics: → maps joint velocities → end-effector velocity → singularities show up when the jacobian collapses in optimization / ML: → gradient flow through layers → backprop is chained jacobians interpretation: → determinant ≠ 0 → transformation is locally invertible → determinant = 0 → information collapse the jacobian is not theory it’s the interface between geometry and computation
Mustafa tweet media
English
18
239
1.5K
42.8K
Boochi 🇳🇬 retweetledi
Francesco Bertolotti
Francesco Bertolotti@f14bertolotti·
This is a simple alternative to low-dimensional embedding methods such as tSNE, UMAP, and PCA. It trains an autoencoder to match the distances of the reference space. The results quite good. 🔗arxiv.org/abs/2603.16568
Francesco Bertolotti tweet mediaFrancesco Bertolotti tweet mediaFrancesco Bertolotti tweet mediaFrancesco Bertolotti tweet media
English
5
80
440
34.5K
Zainab Lawal
Zainab Lawal@Zeeskylaw·
I love rage-baiting my professors in class. Today I asked my CV professor if he thinks, supervised learning is actually just a boring, special case of reinforcement learning where the environment is static and rewards are immediate. He thought I was going insane but we had a nice discussion about evaluative vs. instructive feedback. 😆
English
5
4
48
4.9K
Boochi 🇳🇬 retweetledi
Boochi 🇳🇬
Boochi 🇳🇬@boochi_dot_dev·
But I must also admit that there are certain rare cases where the lines between supervised learning and RL gets blurred. Here’s one good example: In a Contextual Bandit scenario, the agent sees a state (the input), takes an action (the prediction), and gets a reward (the loss signal). If the environment doesn't change based on the agent's action (it's "static") and the reward is given instantly, the boundary between SL and RL becomes incredibly thin. In this case one might argue the Maximum Likelihood Estimation (the heart of supervised learning) is captured within reward function where the reward is 1 for the correct label and 0 for everything else.
English
1
1
0
43
Boochi 🇳🇬
Boochi 🇳🇬@boochi_dot_dev·
There’s more to the RL vs. SFT debate than just evaluative vs. instructive feedback. In most LLM post-training, RL is specifically designed to improve performance “on-policy”. SFT just mimics a static distribution(i.e “off policy”), while RL optimizes the model’s own generated policy against a reward signal (typically while using a KL divergence penalty to ensure it doesn't drift too far from the base model). This greedy reward objective forces the model to explore and master complex strategies that simple supervised imitation can’t capture.
Zainab Lawal@Zeeskylaw

I love rage-baiting my professors in class. Today I asked my CV professor if he thinks, supervised learning is actually just a boring, special case of reinforcement learning where the environment is static and rewards are immediate. He thought I was going insane but we had a nice discussion about evaluative vs. instructive feedback. 😆

English
2
0
1
339
AVB
AVB@neural_avb·
People don’t realize these OpenAI releases aren’t competing with Claude. Anth has no good model at that price range. It’s 5.4 mini vs Gemini 3 Flash It’s 5.4 nano vs Gemini 3.1 Flash Lite Incredible models for narrow tasks that are backbones in most automation workflows.
OpenAI@OpenAI

GPT-5.4 mini is available today in ChatGPT, Codex, and the API. Optimized for coding, computer use, multimodal understanding, and subagents. And it’s 2x faster than GPT-5 mini. openai.com/index/introduc…

English
3
2
13
1.2K
Boochi 🇳🇬
Boochi 🇳🇬@boochi_dot_dev·
“On policy” means the model learns (or is optimised) using its own previous attempts. Exploration makes this possible. “Off policy” means the model is optimised for accurate predictions against a given distribution (eg domain dataset, task, etc). This is what people popularly refer to as “imitation learning”. In theory, a model will achieve better policy generalisation when evaluated on an out of domain dataset through RL than SFT
English
0
0
0
37