Joyce
235 posts

Joyce
@joyceerhl
product @cerebras prev eng @code ✨ opinions mine
San Francisco Katılım Haziran 2019
274 Takip Edilen522 Takipçiler
Joyce retweetledi
Joyce retweetledi
Joyce retweetledi
Joyce retweetledi

@kevinwestmx @Zai_org GLM4.7 on @cerebras is insane. I was impressed with the model performance on my sample test, but I was BLOWN AWAY by the speed. Real window into the future of AI.
English
Joyce retweetledi

Join us next week for Cafe Compute Seattle with the @cerebras, @code, and @github teams!
RSVP: luma.com/cafecompute-se…
English
Joyce retweetledi
Joyce retweetledi
Joyce retweetledi

@OpenAI and @Cerebras have signed a multi-year agreement to deploy 750 megawatts of Cerebras wafer-scale systems to serve OpenAI customers.
This has been a decade in the making.
Deployment begins in early 2026, and when fully rolled out, it will be the largest high-speed AI inference deployment in the world.
OpenAI and Cerebras were both founded in 2015 with radically ambitious goals.
OpenAI set out to build the software that would push AI toward general intelligence.
Cerebras set out to rethink computing hardware from first principles.
Our teams met as far back as 2017. We shared ideas, early work, and a common belief:
there would come a point when model scale and hardware architecture would have to converge.
That point has arrived.
ChatGPT set the direction for the entire industry. It showed the world what AI could be.
Now we’re in the next phase - not proving capability, but delivering it at global scale.
The history of technology is clear on one thing:
speed drives adoption.
The PC industry didn’t operate at kilohertz.
The internet didn’t change the world on dial-up.
AI is no different.
As models grow more capable, speed becomes the bottleneck.
Slow systems limit what users can do, how often they engage, and whether AI becomes infrastructure or remains a novelty.
Cerebras was built for this moment.
By keeping computation and memory on a single wafer-scale processor, we eliminate the data-movement penalties that dominate GPU systems. The result is up to 15× faster inference, without sacrificing model size or accuracy.
That speed changes product design, user behavior, and ultimately productivity.
For consumers, it means AI that feels instantaneous.
For the economy, it means agents that can finally drive serious productivity growth.
For Cerebras, 2026 will be a defining year.
With this collaboration with OpenAI, Cerebras’ wafer-scale technology will reach hundreds of millions - and eventually billions - of users.
We’re proud to work alongside OpenAI to bring fast, frontier AI to people around the world.
This is what a decade of long-term thinking looks like.

English
Joyce retweetledi
Joyce retweetledi
Joyce retweetledi

You can now use GLM-4.7 through Cerebras on AI Gateway.
Cerebras@cerebras
GLM-4.7 from @Zai_org is live on Cerebras! - Frontier intelligence for coding, tool-driven agents, and multi-turn reasoning - Record coding speed: ~1,000 tokens per second (up to 1,700 TPS for other uses) - Strong price-performance: ~10x higher than Sonnet 4.5
English

@spsbuilds @cerebras @Zai_org 👋 the API supports tool calling and structured outputs when used separately, but doesn't currently support having them both in the same request. Learn more: inference-docs.cerebras.ai/capabilities/s…
English
Joyce retweetledi

GLM-4.7 from @Zai_org is live on Cerebras!
- Frontier intelligence for coding, tool-driven agents, and multi-turn reasoning
- Record coding speed: ~1,000 tokens per second (up to 1,700 TPS for other uses)
- Strong price-performance: ~10x higher than Sonnet 4.5
English

@AonSayyed @cerebras @Zai_org 👋 we support 131K context window for GLM 4.7, and these are the full model weights (non-REAPed).
English
Joyce retweetledi

@monstercameron @_MR_WALI_ the best experience I've had with BYOK is @Cerebras and Qwen models in @code :)
English

Hello @pierceboggan 👋🏻
When are we getting some better "0x" models in copilot chat? There are better, cheaper & open-source alternatives, why not provide them?
English












