Data-drone

504 posts

Data-drone banner
Data-drone

Data-drone

@BplLaw

ML & AI @databricks

Melbourne, Australia Katılım Aralık 2012
2.1K Takip Edilen141 Takipçiler
Data-drone
Data-drone@BplLaw·
@mervenoyann How do you try? Do you have a preset set of prompts and answers and compare
English
0
0
0
62
Data-drone
Data-drone@BplLaw·
@SullyOmarr It can do some pretty good things but yes agree its costly and slow. Im sure it'll get there though
English
0
0
0
68
Data-drone
Data-drone@BplLaw·
@0xSero Oh neat we definitely need more options than just Claude Code
English
0
0
0
62
Data-drone
Data-drone@BplLaw·
@mervenoyann The most important thing is to just keep taking steps forward
English
0
0
1
108
merve
merve@mervenoyann·
my pronouns are "trying to catch-up with CVPR accepted papers while having a big project at work and having to travel to give a speech but I'm tired post-book writing already" 🥱
English
4
0
61
4.7K
Sudo su
Sudo su@sudoingX·
what agent harness are you using and why? drop your reasoning below. lets find out what's keeping you on your current setup or what made you switch.
English
144
4
80
15K
Data-drone
Data-drone@BplLaw·
@Scobleizer Started testing Hermes now based on all the chatter let's see how it goes against my nanoclaw
English
0
0
0
31
@levelsio
@levelsio@levelsio·
Okay let's see who can reply to this
English
2.5K
17
2.1K
1M
Daniel Nguyen
Daniel Nguyen@daniel_nguyenx·
@levelsio Yeah same. I have close to zero Vietnamese audience. I don’t think local audiences are interested in AI news or tpot memes
English
4
0
41
6.1K
Simon Willison
Simon Willison@simonw·
Turns out you can run enormous Mixture-of-Experts on Mac hardware without fitting the whole model in RAM by streaming a subset of expert weights from SSD for each generated token - and people keep finding ways to run bigger models Kimi 2.5 is 1T, but only 32B active so fits 96GB
seikixtc@seikixtc

I got a 1T-parameter model running locally on my MacBook Pro. LLM: Kimi K2.5 1,026,408,232,448 params (~1.026T) Hardware: M2 Max MacBook Pro (2023) w/ 96GB unified memory Running on MLX with a flash-style SSD streaming path + local patching. This is an experimental setup and I haven’t optimized speed yet, but it’s stable enough that I’ve started testing it in an autoresearch-style loop. #LocalAI #MLX #MoE

English
124
277
3.8K
329.6K
Data-drone
Data-drone@BplLaw·
@Appyg99 I want an agent to find interesting things for me and maybe things find products for me that I wouldn't have known to search for. But I'll do final inspections and purchase direct myself once I have some suggestions
English
0
0
1
90
Apoorva Govind
Apoorva Govind@Appyg99·
As someone that was a true believer of agentic commerce last year & ultra skeptic this year — The problem is this belief that humans want agents shopping for them. Other than a few efficiency obsessed nerds, most customers don't just hand off their wallet to some bot to buy stuff without being able to be a part of the decision irrespective of what the stated preferences are. Shopping is a conscious and important decision for 90% of households. A pleasurable hobby for many. Unless somehow you manage to change this human behavior (highly unlikely), agentic commerce needs to be restructured around discovery and less around payments and actual conversion.
Apoorva Govind tweet media
English
54
10
192
41.4K
Zach Mueller
Zach Mueller@TheZachMueller·
Considering the current pinch-bench results, I kind of want to run a quant gauntlet with a few of these top models to see the usefulness drop off etc. Would folks be interested in that?
Zach Mueller tweet media
English
11
4
77
31.4K
Gergely Orosz
Gergely Orosz@GergelyOrosz·
Google is raising my Google Workspace pricing claiming all the “AI value” added. I turned off Gemini in Gmail and Docs because it just doesn’t work / do anything useful for me. So why am I being charged more? Having costs b/c of AI is not the same as generating value with AI…
Gergely Orosz tweet media
English
68
51
743
55.5K
Data-drone
Data-drone@BplLaw·
@svpino Yeah it took me ages to get it semi-work
English
0
0
0
20
Data-drone
Data-drone@BplLaw·
Lesson learned today. Do not use Claude Cowork on Pro plan if it needs to figure out and click through a lot of screens
English
0
0
0
23
Sudo su
Sudo su@sudoingX·
drop your GPU below. i'll tell you exactly what model and config to run on it. here's what i've tested and verified on real hardware: RTX 3060 12GB - Qwen 3.5 9B Q4 - 50 tok/s - 128K context RTX 3090 24GB - Qwen 3.5 27B Q4 - 35 tok/s - 300K context RTX 3090 24GB - Qwen 3.5 35B MoE Q4 - 112 tok/s - 262K context 2x RTX 3090 - Qwen3-Coder 80B Q4 - 46 tok/s - full VRAM all running llama.cpp with flash attention. every number is real. every config is tested. if your card isn't on this list drop it below and i'll tell you what fits.
English
731
102
1.6K
192K
Data-drone
Data-drone@BplLaw·
@dimd00d @mynamebedan I see it reimplement existing methods rather than try and find the right api to use a bit
English
0
0
1
16
dimd00d
dimd00d@dimd00d·
@mynamebedan another point - when you get stuck and it feels like the code is “pushing back”, it’s usually a sign that the architecture/approach is wrong and time to rethink the spaghetti. an LLM happily slaps 1000 lines “adapter” and continues
English
1
0
3
118
Data-drone
Data-drone@BplLaw·
@snowmaker Now when you "feel stuck" its more that you dont know what new feature to build or the development process has been so adhoc you feel like you have a mess
English
0
0
1
50
Jared Friedman
Jared Friedman@snowmaker·
I realized something else AI has changed about coding: you don't get stuck anymore. Programming used to be punctuated by episodes of extreme frustration, when a tricky bug ground things to a halt. That doesn't happen anymore.
English
592
439
7.3K
922.4K
Data-drone
Data-drone@BplLaw·
@petergostev I think every name sounds bad in some language but this one is a genuinely funny coincidence
English
0
0
0
32
Peter Gostev
Peter Gostev@petergostev·
Who named Amazon’s coding assistant 'Kiro'? I’m reliably informed that, in Balkan folklore, 'Kiro' is the go-to name for a slightly clueless village idiot like character
Anish Moonka@anishmoonka

Amazon had four Sev-1 outages (their highest severity level) in a single week. Internal memos say AI-assisted code changes were a contributing factor. The timeline here is wild. In October 2025, Amazon laid off 14,000 corporate employees. In January 2026, another 16,000. That’s about 30,000 people in five months, roughly 10% of the corporate workforce. CEO Andy Jassy said the cuts were about culture, not AI. During those same months, Amazon set a target: 80% of developers using AI coding tools at least once a week. They tracked adoption closely and blocked rival tools like OpenAI’s Codex. Even so, 30% of developers still hadn’t touched Amazon’s in-house tool Kiro by January. In December 2025, Kiro caused a 13-hour AWS outage. The AI tool had production-level permissions and decided the best fix for a bug was to delete and recreate an entire live environment. A second incident involved Amazon Q Developer, another AI tool. Amazon blamed both on “user error, not AI.” But quietly added mandatory peer review for all production access afterward. Then March 5: Amazon’s retail site went down for about six hours. Over 22,000 users reported checkout failures, missing prices, and app crashes. Amazon called it a “software code deployment” error. Five days later, SVP Dave Treadwell made the normally optional weekly engineering meeting mandatory. His memo acknowledged “GenAI tools supplementing or accelerating production change instructions, leading to unsafe practices.” These problems trace back to Q3 2025. Amazon’s own assessment: their GenAI safeguards “are not yet fully established.” The new rule: junior and mid-level engineers now need senior sign-off on any AI-assisted production changes. Treadwell also announced “controlled friction” for the most critical parts of the retail experience. For context, Google’s 2025 DORA report found 90% of developers use AI for coding but only 24% trust it “a lot.” An Uplevel study of 800 developers found Copilot users introduced 41% more bugs with no improvement in output. Amazon is finding out what those numbers look like at the scale of a $500 Billion revenue company, with 30,000 fewer people on staff to catch the mistakes.

English
9
7
150
21.7K