stonecobra

8.3K posts

stonecobra

stonecobra

@stonecobra

AI Architect and Engineer

Wasatch Front, UT Katılım Kasım 2008
2.1K Takip Edilen684 Takipçiler
stonecobra
stonecobra@stonecobra·
Love the uptime monitoring with @sentry, when are you going to extend that to make public status pages for a SaaS?
English
0
0
0
28
stonecobra retweetledi
Dillon Mulroy
Dillon Mulroy@dillon_mulroy·
confession: i haven't launched a single sandbox on any provider in 2026. i just use dynamic workers
English
17
5
259
26.3K
stonecobra retweetledi
zach
zach@zachglabman·
I am no oracle but I can say with certainty that manufacturers from my roadtrip noted the need for controls engineers (PLC), technicians, machinists and CAM programmers Manufacturing is tech now. Don’t need to be an ML engineer to feel like you can make an impact
Oregon Territory@OregonTerritory

@zachglabman Great article. As a novice looking for a mid-career change, any advice? Would computer-machine learning be an avenue for entering the sweet spot between tech and manufacturing, or is it going to get washed away by AI?

English
8
3
80
6.5K
stonecobra retweetledi
Gergely Orosz
Gergely Orosz@GergelyOrosz·
Massive divide I’m seeing: A) Startups where the founder hands-on, building with the latest AI tools and best models, sees first-hand what this means and championing everyone to use it, not caring about $$$ B) founder not engaged, devs still think AI (aka Copilot) is “meh”
English
111
100
1.8K
210.3K
stonecobra retweetledi
Benjamin De Kraker
Benjamin De Kraker@BenjaminDEKR·
Wait... so Stripe has Agentic Commerce Protocol (ACP), designed for AI agents. They collaborated with OpenAI on it. But now Shopify has Universal Commerce Protocol (UCP), designed for AI agents They collaborated with Google on it. Both are huge brands. Who wins this?
English
152
40
1.3K
121.4K
stonecobra retweetledi
anshuman
anshuman@athleticKoder·
I rejected a job offer yesterday. Not because of the salary. Not because of the tech stack. Not even because of the long hours they warned me about. But because, when I asked how they evaluate their AI systems, the hiring manager said: "We just ask it some questions and see if the answers sound right." I stared at them for a moment and realized... They just described the biggest problem in AI today. See, "sounds right" isn't a measurement. It's a hope. Here's what proper LLM evaluation actually looks like: - Accuracy: Can it get factual questions right? (Not 80% of the time. Consistently.) - Hallucination rate: How often does it make things up? (This should be near zero for critical applications.) - Bias metrics: Does it treat all groups fairly? (Measured across demographics, not assumed.) Real Evaluation Frameworks: - BLEU scores for translation quality Perplexity for language modeling Human evaluation with inter-annotator agreement Adversarial testing (red teaming) Domain-specific benchmarks (legal, medical, financial) The Process: > Define success criteria BEFORE deployment > Create diverse test sets (not just happy paths) > Measure consistently across model versions > Track performance over time (models drift) Have humans validate edge cases Why This Matters: Before proper evals: "Our model is amazing!" (based on cherry-picked examples) After proper evals: "Our AI achieves 94.2% accuracy on domain X, with known failure modes Y and Z" The difference? One builds trust. The other destroys it when reality hits. The kicker: Most companies are still in the "sounds right" phase. They're deploying models evaluated by vibes, not metrics. Just like you wouldn't join a team that deploys code without tests, you shouldn't join one that deploys AI without proper evaluation. What's your experience with LLM evaluation? Are we measuring what actually matters?
English
40
42
559
137.1K
stonecobra
stonecobra@stonecobra·
@arimorcos This is gonna be huge for a lot of embedding use cases
English
0
0
0
144
Ari Morcos
Ari Morcos@arimorcos·
Curating petabytes of data requires ultrafast embedding of massive amounts of data. To do so, we've been using our own luxical embeddings, which are dramatically faster than neural embeddings and run on CPU. Today, we're releasing Luxical so everyone can benefit!
Ari Morcos tweet media
Luke Merrick@lukemerrick_

Just dropped a new text embedding methodology. Fast as heck on CPU only and still great for document similarity analysis, clustering, and classification. How? Use a tiny ReLU network to approximate a big transformer from lexical (term frequency / bag of words) features.

English
13
66
602
54.2K
Noel Ceta
Noel Ceta@noelcetaSEO·
2/ The caching layers Layer 1: Browser cache - User's browser caches images, CSS, JS - Cache-Control headers Layer 2: CDN cache (Cloudflare) - Edge servers cache entire pages - Serves cached content globally Layer 3: Server cache (Redis/Memcached) - Caches database queries - Caches generated HTML Layer 4: Database cache - MySQL query cache - Caches frequent queries Each layer compounds the effect.
English
2
0
1
177
Noel Ceta
Noel Ceta@noelcetaSEO·
Caching = storing generated pages for reuse. Everyone thinks caching is just for speed. But proper caching directly impacts SEO: - Faster responses = more crawling - Lower server load = better uptime - Better Core Web Vitals = higher rankings Client's caching strategy: Traffic increased 67% in 5 weeks. Here's the complete caching playbook: 🧵👇
English
11
6
46
4.2K
stonecobra retweetledi
Gergely Orosz
Gergely Orosz@GergelyOrosz·
Imagine if all major tech companies took outages as seriously, and held themselves accountable publicly as Cloudflare does. But 99% of them do not: and so I trust Cloudflare more than almost any other company (including the hyperscalers). Cloudflare next level in transparency
Matthew Prince 🌥@eastdakota

We let the Internet down today. Here’s our technical post mortem on what happened. On behalf of the entire @Cloudflare team, I’m sorry. blog.cloudflare.com/18-november-20…

English
21
54
903
76.3K
stonecobra retweetledi
Matt Silverlock 🐀
Matt Silverlock 🐀@elithrar·
It amazes me after all these years that Azure Blob doesn't have an S3 compatible API. I can't begin to imagine how many customers must've asked, what architectural decisions prevent it, and the millions of collective hours spent supporting its bespoke API.
English
14
5
216
15.5K
stonecobra retweetledi
Andrew Côté
Andrew Côté@Andercot·
It is TWICE as expensive to install a megawatt of nuclear reactor on US soil than it is to put a megawatt of nuclear power into a US Aircraft Carrier. Why? The US Navy doesn't have to go through the NRC.
Andrew Côté@Andercot

Honestly, the US Navy should just become America's nuclear power provider. US Navy Cost: $2 billion for 2 x 400 MW reactors in Ford-class aircraft carrier NuScale: $10 bn for 500 MW reactor Westinghouse: ~$8 bn for 1000MW reactor

English
243
1.6K
17.1K
786.5K
stonecobra retweetledi
Jaana Dogan ヤナ ドガン
Lately I’ve been feeling depressed because decades of our hard work is completely gone like it never existed. I heard from others that they also find it very hard to dial into the new norm of low quality software engineering.
English
152
116
2.8K
743.6K
stonecobra retweetledi
Jimmy Song (송재준)
Jimmy Song (송재준)@jimmysong·
The minority opinion is almost always better informed and better thought through, just by the nature of it being a minority opinion. - log p is pretty descriptive that way.
English
8
4
57
7.4K