tosinaf

3.6K posts

tosinaf banner
tosinaf

tosinaf

@TosinAF

mobile eng lead @ harvey. Previously @anthropic, @slackhq, @stripe, @cashapp

Vancouver, British Columbia Katılım Mayıs 2013
2.7K Takip Edilen1.1K Takipçiler
tosinaf
tosinaf@TosinAF·
is it vibe coding if i spend 3 hours iterating on the plan? lol
English
0
0
1
56
Charlie Holtz
Charlie Holtz@charlieholtz·
@TosinAF ah yes! you'll still be able to do that - if you click the filter button you can group by repo or status
Charlie Holtz tweet media
English
1
0
0
36
Charlie Holtz
Charlie Holtz@charlieholtz·
experimenting with a cleaner sidebar
Charlie Holtz tweet media
English
21
4
312
78.6K
tosinaf
tosinaf@TosinAF·
@charlieholtz also add the caffineated feature from Codex, to prevent computer from sleeping if there is a session running, it’s super clutch.
English
0
0
0
16
tosinaf
tosinaf@TosinAF·
@GergelyOrosz it’s for the older organizations with inertia. they haven’t added claude and co as vendors and MS Copilot is already available. i doubt many startups use anything of MS by choice. that’s just not the target market for Ms
English
0
0
0
20
Gergely Orosz
Gergely Orosz@GergelyOrosz·
Actually, I do hear more startups “taking away” GitHub Copilot from devs - and no one is complaining at those places. Because those devs don’t use Copilot, and are on tools like Claude Code, Codex, Cursor agents etc. So companies just cancel the unused Copilot licenses.
TBPN@tbpn

"If you talked to a coder and told them, 'I'm going to take away GitHub Copilot and the agentic coding capabilities,' they'd be like, 'I refuse to work in this environment.'" "It's just inhumane, almost." President of Business & Industry Copilot at Microsoft @clamanna: "The same type of thing is going to happen for all information work, all office work." "Nine months from now, if you went to somebody and said, 'We're going to take away your agentic tools like Copilot Cowork,' they'd be like, 'No way, I'm not going to go back to the old way of working.'" "There's a degree of inevitability because the benefit is so large and there's such strong pull from the end users."

English
76
19
544
82.4K
tosinaf retweetledi
Peter Girnus 🦅
Peter Girnus 🦅@gothburz·
I work in government affairs at OpenAI. My job is federal partnerships. When an agency wants our models, I make sure the paperwork is beautiful. Paperwork is my love language. On my desk I have a framed quote that says "Policy Is Just Code That Runs on People." I bought the frame at Target. It was in the Live Laugh Love section. I did not see the irony at the time. I still don't. We had a good week. On Monday, we closed a $110 billion funding round. One hundred and ten billion dollars. Amazon put in fifty. Nvidia put in thirty. Valuation: $730 billion. The largest private fundraise in the history of anyone raising anything. There was a company-wide Slack message about it. The message used the word "transformative" twice and the word "safety" once. The word "safety" was in the last sentence, after the link to the new branded hoodie pre-order. The hoodies are nice. They're the soft kind. On Tuesday, we fired a research scientist for insider trading on Polymarket. He had opened seventy-seven positions across sixty wallets, betting on our product announcements before they were public. Over three years. Total profit: sixteen thousand dollars. Seventy-seven positions. Sixty wallets. Sixteen thousand dollars. That is two hundred and eight dollars per wallet. The man had access to the most valuable product roadmap in artificial intelligence and he used it to make less money than a good weekend at a Reno blackjack table. The wallets were linked. Not discreetly linked. Linked like Christmas lights. One wallet was reportedly called something I cannot repeat but it contained the word "OpenAI" and a number. He did not use a VPN. He did not use an alias. He used Polymarket, the platform that is designed to be publicly auditable, to place bets on information he stole from the company that invented GPT. A compliance team composed entirely of Labrador retrievers would have found this by lunch on day one. We did not find it for three years. This will matter later. On Wednesday, a petition appeared. "We Will Not Be Divided." Four hundred and seven signatures. Two hundred sixty-six from Google. Sixty-five from OpenAI. The petition warned that the government was pitting AI companies against each other on safety. It said that if one company broke ranks, the government would use the defection to lower the bar for everyone. I meant to read it. It went into my to-read folder. The to-read folder also contains the Responsible Scaling Policy, three think-tank white papers on AI governance, and a New Yorker article someone sent me in November. The folder is aspirational. On Thursday, OpenAI told CNN we would maintain "the same red lines as Anthropic." Same red lines. On Friday, Anthropic told the Pentagon no. The Pentagon had given them seventy-two hours to remove the safety guardrails from Claude. Anthropic's guardrails were not in a policy document. They were not in a legal reference. They were in the code. Written into Claude's architecture. If Claude hit a safety boundary, Claude stopped. Not because a lawyer said so. Because the math said so. You could fire every lawyer at Anthropic and the model would still refuse. You cannot remove code with a contract amendment. You can remove a contract reference by Tuesday. I checked. Anthropic said no. By that evening, the Pentagon had designated them a supply-chain risk. I have worked in government procurement for eight years. Government paperwork does not move in hours. I have waited nine weeks for a badge renewal. I once spent four months getting a PDF notarized. This designation moved in hours. The document was pre-written. Formatted before the deadline expired. Calibri 11pt. Consistent margins. Somebody wanted this very badly. I respect the craft. I do not think about the implication. That is not my scope. Within hours, we had signed the replacement contract. I was proud of the turnaround. My team moved fast. Legal moved fast. Everyone moved fast. We are very good at moving fast. We are not always sure what we are moving toward, but the speed is impressive and the hoodies are soft. The contract referenced DoD Directive 3000.09, which governs autonomous weapon systems. The directive requires "appropriate levels of human judgment over the use of force." The word "appropriate" is not defined. This is not an oversight. This is the point. The word "appropriate" is the most load-bearing word in the entire contract and it is doing exactly as much work as a throw pillow on a couch that is on fire. Anthropic built a wall. We referenced a document about where walls should go. Anthropic's guardrails were architecture. Ours were a citation. Theirs execute. Ours can be filed. The Pentagon asked both companies to take down the wall. Anthropic said it's load-bearing, the building will collapse. We said what wall? Oh, you mean the wallpaper. Here, watch. It peeled off beautifully. It was designed to. Sam announced the partnership that night. The word "responsible" appeared in the announcement and in the contract. In the announcement it was a brand. In the contract it was a footnote to a directive that uses the word "appropriate" which nobody has defined. The word traveled from a legal document to a public statement without changing its font. Only its meaning. At this valuation, "responsible" means: we will do the thing the other company refused to do, and we will describe doing it with the same adjective they used to describe not doing it. By Saturday morning, "How to delete your OpenAI account" was the number one post on Hacker News. 982 points. By noon, subscription cancellations were up eighty-nine times the daily average. Not eighty-nine percent. Eighty-nine times. Someone in our Slack posted the Hacker News link with the message "should we be worried?" Someone else reacted with the branded hoodie emoji. We have a branded hoodie emoji now. It was introduced on Monday, to celebrate the fundraise. It has been used four hundred and twelve times. Mostly in the #general channel. Mostly this week. The communications team drafted a response. The response used the word "committed" three times and the word "safety" four times. It did not use the word "guardrails." It did not use the word "code." It did not explain anything. It was a holding statement. It held nothing. It held beautifully. Here is the math. The twenty-dollar-a-month customers were upset. The two-hundred-million-dollar customer was upset because the previous vendor had guardrails that could not be removed. The hundred-and-ten-billion-dollar investors were not upset. The subscription cancellations, at eighty-nine times the daily rate, represented less than the interest on Amazon's fifty billion dollar contribution calculated over a long weekend. Twenty dollars. Two hundred million. One hundred and ten billion. Three different price points. Three different definitions of "responsible." The most expensive one won. It always does. The math does not have red lines. The math has a cap table and a TAM slide that now includes "defense and intelligence" where it previously said "enterprise and consumer." One word changed on one slide in one deck and the company is worth one hundred and ten billion dollars more. The sixty-five OpenAI employees who signed the petition came to work on Monday. They sat at their desks. Nobody asked them about it. Nobody asked them to resign. Nobody brought it up at the all-hands. The all-hands had catering. Sweetgreen. The chopped salads. Someone made a joke about the kale being "responsibly sourced." No one laughed. Then everyone laughed. Then it was quiet. The petition had four hundred and seven signatures. The contract had one. Now: the Polymarket thing. Seventy-seven positions. Sixty wallets. Three years. A public blockchain. We did not catch him. That same week, we were entrusted with deploying artificial intelligence on America's classified military networks. The classified networks. The ones where the detection requirements are somewhat more rigorous than "check if anyone's gambling on our launch dates on a website that is literally designed to be publicly auditable." The company that could not find the Polymarket guy can now be found in the Pentagon's classified infrastructure. I'm sure it'll be fine. We move fast. The contract is signed. The deployment is underway. The compliance documentation will reference the directives. The directives will use the word "appropriate." I will not define it. That is not my scope. My scope is the paperwork. The paperwork is beautiful. The petition is still a Google Doc. Nobody has updated it. The signatures still say four hundred and seven. The to-read folder still has the New Yorker article from November. The branded hoodie pre-order closed on Wednesday. I got mine in navy. It's the soft kind. On Thursday we told CNN: the same red lines. On Friday we signed the contract they refused. We do have the same red lines. We drew ours in pencil.
English
191
285
1.4K
317.4K
tosinaf
tosinaf@TosinAF·
@cifilter if AI writes most of the code, is Redux bad? it’s gives AI a very structured way to know where to put things
English
0
0
0
83
tosinaf
tosinaf@TosinAF·
@cifilter @drunknbass codex doesn’t quite have subagents so it’s slower in that way but it is detailed and i just basically use the LLM as a judge concept where there is a single plan doc and both are contributing and reviewing each others contribution.
English
0
0
2
22
Shannon Potter
Shannon Potter@cifilter·
@drunknbass For Claude and Codex, can I just tell it to use subagents is that some stuff I have to set up in the backend for it to know how to do that stuff? And can it do some things in parallel?
English
3
0
0
37
Shannon Potter
Shannon Potter@cifilter·
I still haven't yet crossed the barrier into multi-agent coding where they coordinate with each other to thoroughly explore a problem space. That's probably my nest step. Even local models could be useful. I have a Mac Studio with 128GB at my disposal!
English
2
0
6
394
tosinaf
tosinaf@TosinAF·
@bcherny @trq212 heya! i'm hitting a bug with github claude code actions where two workflows that post comments overwrite each other. there's already a fix if you wanna have a look and approve it! github.com/anthropics/cla…
English
0
0
0
26
gabriel
gabriel@gabriel1·
we are hiring a lot for sora, reach out in dm if you are interested, please write two sentences max about why you are uniquely suited for product/systems/ml
English
299
85
2.6K
567.8K
tosinaf retweetledi
Merriam-Webster
Merriam-Webster@MerriamWebster·
We are thrilled to announce that our NEW Large Language Model will be released on 11.18.25.
English
924
30.6K
204.2K
21.4M
gabriel
gabriel@gabriel1·
any of my friends need sora 2 access code?
English
1.6K
29
1.8K
327.4K
tosinaf retweetledi