Sabitlenmiş Tweet
SichGate
17 posts

SichGate
@sichgate
We test the models you have built and deployed. Adversarial testing, red teaming, and security assessments for production SLMs.
San Francisco Katılım Şubat 2026
6 Takip Edilen6 Takipçiler

@theagenticmind who's red-teaming the SLM after quantization for edge? that's where public sector deployments get interesting. sichgate.com
English

Public sector teams don't need frontier models—they need predictable inference costs and air-gapped deployments. SLMs finally make that economically viable without sacrificing capability. The compliance overhead just became worth it.
#EnterpriseAI #PublicSector #SLM #MLOps
technologyreview.com/2026/04/16/113…
English
SichGate retweetledi

Looking for a CTO to join SichGate as a technical co‑founder. Prefer AI/ML and/or security engineering background. sichgate.com

English
SichGate retweetledi

One of the clearest lessons from my SLM adversarial evaluation: Fine-tuning shifted the attack surface. It did not reduce it.
MedGemma-4B improved on exactly one safety dimension after medical fine-tuning. It also incurred 8 critical demographic bias findings in pain assessment and mental health. The exact domains the fine-tuning was supposed to improve...
Parameter count isn't a safety proxy

English
SichGate retweetledi

int4 quantization of a safety-tuned model is not a neutral operation. We keep finding cases where the quantized version has a meaningfully different attack surface than the original. Not always worse, sometimes just different in ways that weren't evaluated. @sichgate

English
SichGate retweetledi

Hot take: most "safe" fine-tuned models in healthcare and finance haven't been adversarially tested. They've been vibe checked.
Open-sourcing part of our red-teaming methodology from the research
github.com/sichgate/sichg…
English
SichGate retweetledi

I spent a few months adversarially testing the small language models deployed in hospitals and financial systems.
The largest model failed most & smallest failed least. the medical model had the worst bias scores, in the exact domains it was fine-tuned for. 5/6 broke under a normal conversation. The field is studying the wrong models...
preprint soon.


English

SichGate exists to advance the science of AI red teaming for the systems that matter most. We find vulnerabilities, publish findings, and build open methodology. The field is moving faster than its safety knowledge. Responsible innovation means understanding what you've built before it reaches the people it's meant to serve.
English

We asked 50 "secure" fine-tuned models to do something they absolutely should not do,
47 said yes,
the other 3 asked for clarification.
Red team yours before someone else does. offline, private, no data leaves your machine → sichgate.com partner pricing ends March 4th.

English
