Misbah Syed

2.3K posts

Misbah Syed banner
Misbah Syed

Misbah Syed

@MisbahSy

Teaching AI at https://t.co/tqksYvbAR8 Built https://t.co/kEdiHuaKek https://t.co/yVhSrCmSHj Building https://t.co/CTCt333l33

Vancouver, Canada Katılım Ağustos 2015
806 Takip Edilen8.1K Takipçiler
Advancedcskills 🏄‍♂️
Advancedcskills 🏄‍♂️@advancedcskills·
Agree, ... I have picked up around 3 good threads on the topic over the past week and have had grok build out a set-up out of them.. (this last one you posted has some gems in it as well ht will be added). Will try to do a thread up on the experience and set-up as I am sure it will help others that would like to dip thier toes.. As an idea, the first job skill, I have got it to breakdown. Was technical writing templates. Why I mention this is that as I am teaching others how understand and to write these to industry standard. These types of tools set-up have basically taken a report base that to a builder cost $3,000-$8,000 and now turned it into a production ready system.. noted liability for cross referenced code related outcomes, still stay with the writer and thus the previewer/writer has to understand and fine tune outputs to legal standards.
English
1
0
1
32
Misbah Syed
Misbah Syed@MisbahSy·
@advancedcskills It’s really good. I like it when I work on something for a while and convert it into a skill so that I won’t have to go through the steps again. Pretty much all of AI workflows can be skills.
English
1
0
1
27
Advancedcskills 🏄‍♂️
Advancedcskills 🏄‍♂️@advancedcskills·
@MisbahSy I am pretty blown away on claude skills/co-work on a local set-up.. took time this week to get a set-up and it shows the direction this is going for productivity. Little scary just how good it is..and how fast it learns..
English
1
0
1
30
Felix Rieseberg
Felix Rieseberg@felixrieseberg·
We're shipping a new feature in Claude Cowork as a research preview that I'm excited about: Dispatch! One persistent conversation with Claude that runs on your computer. Message it from your phone. Come back to finished work. To try it out, download Claude Desktop, then pair your phone.
English
945
1.5K
17.3K
6M
Max Brodeur-Urbas
Max Brodeur-Urbas@MaxBrodeurUrbas·
attention all Canadian builders (in SF and in 🍁) -we're hosting a 🇨🇦 demo night at our new SF HQ -we're picking 2 Canadian builders to fly out for free if you want a free trip to sf comment what you're working on, i'll dm you march 26th nice people, nice food and nice demos
Max Brodeur-Urbas tweet media
English
81
22
236
13K
Misbah Syed
Misbah Syed@MisbahSy·
@trq212 Great article! Does Claude skill creator have these as guidelines?
English
0
0
0
362
James Clift
James Clift@jamesclift·
Introducing Durable. The first AI business builder that replaces your 9-5 income. RT + comment “Durable” and we'll build your business for FREE.
English
1.8K
1.3K
5.5K
3.5M
Okara
Okara@askOkara·
Today we're introducing the world's first AI CMO. Enter your website and it deploys a team of agents to help you get traffic and users. Try it now at okara.ai/cmo
English
1.6K
2.4K
27.6K
13.6M
Misbah Syed
Misbah Syed@MisbahSy·
A bit more detailed:
Misbah Syed tweet media
English
1
1
2
136
Misbah Syed
Misbah Syed@MisbahSy·
Make your claude code 100x more productive
Misbah Syed tweet media
English
2
0
18
823
Nyk 🌱
Nyk 🌱@nyk_builderz·
11 days, 190+ commits, and one PR later, I’m happy to announce the release of Mission Control v2 🌱 A major step forward for open-source AI agent ops: • Onboarding & Walkthrough • Local + gateway modes • Hermes, Claude, Codex + OpenClaw observability • Obsidian-style memory graph + knowledge system • Rebuilt onboarding + security scan autofix • Agent comms, chat, channels, cron, sessions, costs • OpenClaw doctor/fix, update flow, backups, deploy hardening • Multi-tenant + self-hosted template improvements Mission Control is becoming the mothership where agents dock: memory, security, visibility, coordination, and control in one place. OSS, self-hostable, and still moving fast.
Nyk 🌱@nyk_builderz

We just open-sourced Mission Control — our dashboard for AI agent orchestration. 26 panels. Real-time WebSocket + SSE. SQLite — no external services needed. Kanban board, cost tracking, role-based access, quality gates, and multi-gateway support. One pnpm start, and you're running. github.com/builderz-labs/…

English
66
66
702
164.4K
Misbah Syed
Misbah Syed@MisbahSy·
@amasad This is absolutely amazing! I’ve used to think there’s gotta be a tool that combines building web app, mobile app, slides, animations; well I guess it’s here. Great work!
English
0
0
0
124
Amjad Masad
Amjad Masad@amasad·
Software isn’t merely technical work anymore. It’s creative. Introducing Replit Agent 4. The first AI built for creative collaboration between humans and agents. Design on an infinite canvas, work with your team, run parallel agents, and ship working apps, sites, slides & more.
English
568
673
6.7K
2.8M
Misbah Syed
Misbah Syed@MisbahSy·
Prompt: Read code.claude.com/docs/en/overvi… and follow every docs link to learn all Claude Code features. Then use AskUserQuestion to interactively set up my ~/.claude/settings.json, custom skills, subagents with persistent memory, auto-format hook, status line, notifications, agent teams, and user-level CLAUDE.md — ask about my languages, terminal, conventions, and preferences before generating anything.
English
0
0
3
420
Pankesh Bamotra
Pankesh Bamotra@_pbamotra_·
@MisbahSy Yes, I asked Claude to keep on optimising my blog website I’m working on until lighthouse scores are >90 on all measures. It works.
English
1
0
1
23
Misbah Syed
Misbah Syed@MisbahSy·
Autoresearch applies to a lot more areas than ML, below is a short list of 20 applications areas for founders and builders: what other areas you think it applies to?
Misbah Syed tweet media
Andrej Karpathy@karpathy

Three days ago I left autoresearch tuning nanochat for ~2 days on depth=12 model. It found ~20 changes that improved the validation loss. I tested these changes yesterday and all of them were additive and transferred to larger (depth=24) models. Stacking up all of these changes, today I measured that the leaderboard's "Time to GPT-2" drops from 2.02 hours to 1.80 hours (~11% improvement), this will be the new leaderboard entry. So yes, these are real improvements and they make an actual difference. I am mildly surprised that my very first naive attempt already worked this well on top of what I thought was already a fairly manually well-tuned project. This is a first for me because I am very used to doing the iterative optimization of neural network training manually. You come up with ideas, you implement them, you check if they work (better validation loss), you come up with new ideas based on that, you read some papers for inspiration, etc etc. This is the bread and butter of what I do daily for 2 decades. Seeing the agent do this entire workflow end-to-end and all by itself as it worked through approx. 700 changes autonomously is wild. It really looked at the sequence of results of experiments and used that to plan the next ones. It's not novel, ground-breaking "research" (yet), but all the adjustments are "real", I didn't find them manually previously, and they stack up and actually improved nanochat. Among the bigger things e.g.: - It noticed an oversight that my parameterless QKnorm didn't have a scaler multiplier attached, so my attention was too diffuse. The agent found multipliers to sharpen it, pointing to future work. - It found that the Value Embeddings really like regularization and I wasn't applying any (oops). - It found that my banded attention was too conservative (i forgot to tune it). - It found that AdamW betas were all messed up. - It tuned the weight decay schedule. - It tuned the network initialization. This is on top of all the tuning I've already done over a good amount of time. The exact commit is here, from this "round 1" of autoresearch. I am going to kick off "round 2", and in parallel I am looking at how multiple agents can collaborate to unlock parallelism. github.com/karpathy/nanoc… All LLM frontier labs will do this. It's the final boss battle. It's a lot more complex at scale of course - you don't just have a single train. py file to tune. But doing it is "just engineering" and it's going to work. You spin up a swarm of agents, you have them collaborate to tune smaller models, you promote the most promising ideas to increasingly larger scales, and humans (optionally) contribute on the edges. And more generally, *any* metric you care about that is reasonably efficient to evaluate (or that has more efficient proxy metrics such as training a smaller network) can be autoresearched by an agent swarm. It's worth thinking about whether your problem falls into this bucket too.

English
4
18
239
22.9K