clovy

1.7K posts

clovy banner
clovy

clovy

@realclovisworld

Artificial intelligence & machine learning .... || python 🐍 || developer 💻 || Trader 📉 || Athlete ⚾️|| student 👨🏾‍🎓

Kampala, Uganda Katılım Mart 2020
1K Takip Edilen969 Takipçiler
clovy
clovy@realclovisworld·
@CirylCan Hey, let’s connect 👍
English
0
0
1
4
Ciryl
Ciryl@CirylCan·
I m looking to #connect with builders and learners in tech whether you re just starting out or already deep in the field. • Frontend • Backend • Full-stack • DevOps • Leetcode prep • AI / ML • Data Science • Freelancing • Startups say hi and connect!
English
66
1
46
1.5K
prince
prince@prince_twets·
Hey Twitter 👋 SWE | Building AI systems & indie SaaS Looking to #connect with people into: → System Design → Backend Dev → AI/ML Learning → Open Source → Tech Jobs India What are you building? 👇 Follow = follow back ✅ #LetsConnect #100DaysOfCode
English
52
1
36
1.6K
Sarthak
Sarthak@Sarthak4Alpha·
If you’re in tech, let’s connect. 🤝
English
107
2
60
2.6K
Tanuj
Tanuj@tanujDE3180·
Dear @X algorithm, please #connect me with these people - Software Engineering - Data Engineering - Machine Learning - Gen AI & Agentic AI - Frontend - Backend - Full-stack - SaaS - Data Scientist Let's connect 👨‍💻🔥
English
53
1
49
1.5K
Cat3lyst
Cat3lyst@cat3lyst·
If you're in tech. Let's connect
English
53
0
57
1.6K
Mira
Mira@the_Fitcode·
if you are in Tech let's connect and engage 📈🎀🎀
English
479
13
534
21.6K
clovy
clovy@realclovisworld·
@oprydai True, as we are leaning towards archiving AGI. I think neuroscience comes in clutch to deeply understand how the human brain works ✅
English
0
0
0
10
Mustafa
Mustafa@oprydai·
AI without neuroscience is blind pattern fitting you’re copying outputs without understanding the system that inspired it study neuroscience what you unlock: → how real intelligence encodes information → how learning actually happens (plasticity, not just gradients) → how memory is structured (not just vectors, but dynamics) key parallels: → neurons ≠ artificial neurons real ones spike, adapt, rewire → learning ≠ backprop brains use local rules, timing, chemistry → intelligence ≠ scale it’s efficiency, structure, constraints why it matters: in deep learning: → architectures come from biology (CNNs, attention roots in perception) in AGI: → you need models of cognition, not just bigger models in robotics: → perception + action loops are biological problems in optimization: → energy-based views come straight from brain dynamics what to study: → spiking neural networks → synaptic plasticity (hebbian learning, STDP) → predictive coding → cortical hierarchies direction: → current AI = approximation → neuroscience = source code if you ignore the source you plateau fast
Mustafa tweet media
English
35
38
211
8K
clovy
clovy@realclovisworld·
@embedrapp Wow Thank you , this is a very good tool I’m definitely going to try it out 👍
English
0
0
1
21
Embedr
Embedr@embedrapp·
@realclovisworld the classic breadboard spaghetti phase. love it. if you want to skip the 'debugging pin definitions for 3 hours' part next time, give Embedr a try. it's an AI IDE we built to make this exact workflow way faster: embedr.app
English
1
0
4
46
clovy
clovy@realclovisworld·
Wires everywhere, LEDs blinking, and it actually works 😅 Built an Arduino-based ultrasonic distance sensor with live display output. There’s something satisfying about turning raw components into a working system. #BuildInPublic #Arduino #Hardware #Tech #IOT
clovy tweet media
English
3
0
4
98
clovy
clovy@realclovisworld·
@realthomasgu 😂you’re definitely getting kicked out of the forest too
English
0
0
1
10
Thomas Guthrie
Thomas Guthrie@realthomasgu·
Got kicked out of the house for coding all the time so I moved my setup to the forest 📈
Thomas Guthrie tweet media
English
96
9
252
6.2K
Sai
Sai@yuj_unite·
If you are in tech Let's connect and grow together
English
71
0
54
2.9K
Joy M.
Joy M.@femininow·
If you're in data let's connect
English
22
5
73
2.4K
clovy
clovy@realclovisworld·
Any recommendations to IOT resources to enhance my knowledge on the understanding of working with hardware …😁
English
0
0
0
34
Anaya 🌸
Anaya 🌸@Anaya_sharma876·
Be Honest , as a coder which one do you prefer ?
Anaya 🌸 tweet media
English
68
3
194
4.9K
clovy
clovy@realclovisworld·
Well explained 👏🏾
Dhanian 🗯️@e_opore

TRANSFORMER ARCHITECTURE IN LLMs Large Language Models such as GPT, LLaMA, and PaLM are powered by a neural network design known as the Transformer Architecture. WHAT IS A TRANSFORMER? → A Transformer is a deep learning architecture designed for sequence processing → It analyzes relationships between words in a sentence simultaneously → Unlike older models, it processes all tokens in parallel This enables: → Faster training → Better contextual understanding → Scalability to billions of parameters WHY TRANSFORMERS REVOLUTIONIZED AI Before Transformers, models relied on: → Recurrent Neural Networks (RNNs) → Long Short-Term Memory networks (LSTMs) These had limitations: → Slow sequential processing → Difficulty handling long-range dependencies Transformers solved these issues using attention mechanisms. CORE COMPONENTS OF TRANSFORMER ARCHITECTURE 1) TOKEN EMBEDDINGS → Text is first broken into tokens → Tokens are converted into numerical vectors called embeddings → These vectors capture semantic meaning Example: → "cat" and "kitten" produce similar vectors → "car" and "engine" also have related embeddings 2) POSITIONAL ENCODING Transformers process tokens in parallel, so they need positional information. → Positional encoding adds information about word order → Helps the model understand sentence structure Example: → "Dog bites man" → "Man bites dog" Word order changes meaning, and positional encoding helps capture this. 3) SELF-ATTENTION MECHANISM Self-attention is the core innovation of the Transformer. → Each token looks at every other token → The model calculates how important each word is to another Example sentence: "The animal didn't cross the street because it was tired." The model learns that "it" refers to "animal", not "street". 4) QUERY, KEY, AND VALUE MATRICES Self-attention works using three vectors: → Query (Q) – What the word is asking about → Key (K) – What the word represents → Value (V) – The information carried by the word Attention scores determine which words influence others. 5) MULTI-HEAD ATTENTION Instead of one attention mechanism, Transformers use multiple attention heads. This allows the model to capture different relationships simultaneously. Example: → One head learns grammar → Another learns semantic meaning → Another captures long-distance relationships 6) FEED-FORWARD NEURAL NETWORK After attention, each token passes through a feed-forward neural network. This layer: → Applies nonlinear transformations → Learns deeper feature combinations → Refines token representations 7) LAYER NORMALIZATION AND RESIDUAL CONNECTIONS These help stabilize deep networks. → Residual connections allow gradients to flow better → Layer normalization stabilizes training Together they enable very deep Transformer models. TRANSFORMER LAYER STACKING LLMs stack dozens or even hundreds of Transformer layers. Example: → GPT-3 has 96 Transformer layers → Each layer refines contextual understanding Process flow: → Tokens → Embeddings → Attention → Feedforward → Next Layer → Output OUTPUT GENERATION At the final layer: → The model predicts the probability of the next token → The highest probability token is selected → The process repeats to generate text This is how LLMs produce coherent responses. WHY TRANSFORMERS ARE IDEAL FOR LLMs Transformers enable: → Long-context understanding → Parallel computation → Massive scalability → High-quality text generation They are now the backbone of: → ChatGPT-style assistants → AI coding tools → Document summarization systems → AI search engines QUICK NOTE Understanding Transformer architecture is essential for anyone building modern AI systems and LLM-powered applications. Grab the LLM ENGINEERING HANDBOOK: codewithdhanian.gumroad.com/l/haeit

English
0
0
2
65
Eric Liu
Eric Liu@ericliujt·
Got accepted into YC
Eric Liu tweet media
English
414
49
2.6K
70.5K
clovy retweetledi
ℏεsam
ℏεsam@Hesamation·
what an underrated life skill nobody taught us about. you can achieve ANYTHING by being delusional. strategic delusion isn’t denial, but amplification of beliefs you want to see turn into reality one day. act on that delusion long enough until it’s no longer a dream.
ℏεsam tweet media
English
11
86
917
20.6K
Ann maina
Ann maina@nyagoh_·
Somehow a babe ended up in IoT.
Ann maina tweet media
English
28
23
295
4.5K
kritika
kritika@vibeonX69·
As a developer, What will you watch first if someone give you this setup
kritika tweet media
English
70
3
122
4.8K