H2LooP

39 posts

H2LooP banner
H2LooP

H2LooP

@h2loopai

H2Loop is an AI lab building domain-specific intelligence for lower-level system software and enterprise infrastructure.

Bengaluru Se unió Ağustos 2024
3 Siguiendo18 Seguidores
H2LooP
H2LooP@h2loopai·
Introducing H2LooP Spark: the first domain-specialized autocomplete model for embedded software. A 7B model that beats Claude Opus 4.6 and Qwen3-Coder-30B on embedded code completion. Not fine-tuned. Continually pre-trained on 23B tokens of firmware, datasheets, and vendor SDKs
H2LooP tweet media
English
4
3
7
333
H2LooP
H2LooP@h2loopai·
We built SpecMap: an agentic pipeline that maps vendor datasheets directly to code symbols, across 13 embedded domains 100B raw tokens curated down to 23B. The result: a model that knows the exact register offset, the exact intrinsic opcode, and the exact pin mapping.
English
0
0
2
39
H2LooP
H2LooP@h2loopai·
General LLMs fail at embedded code because - Infineon TriCore intrinsics - NXP eDMA scatter/gather docs, and - AURIX ATOM timer pin maps simply don't exist in standard pre-training data.
English
0
0
2
36
H2LooP
H2LooP@h2loopai·
Token accuracy on held-out embedded code (13 domains, 9 repos never seen during training): → H2LooP Spark 7B: 34.1% → Qwen3-Coder-30B: 24.6% → Claude Opus 4.6: 24.1% → Base OLMo-7B: 16.8% +108% over base. Leads frontier models that are 4–50x larger.
English
0
0
1
172
H2LooP
H2LooP@h2loopai·
MISRA compliance is mandatory in safety-critical C. We built a compact SLM to automate it. Then benchmarked against frontier models 100x its size.
H2LooP tweet media
English
5
2
4
85
H2LooP
H2LooP@h2loopai·
Runs locally. No code leaves your environment. Domain-specialized SLMs matching frontier models at 1/1000th the size.
English
0
0
0
15
H2LooP
H2LooP@h2loopai·
Rule family performance: Pointer safety (11.x, 18.x): 100% fix rate Control flow (15.x): matches Gemini Pro, fewer edits Initialization (9.x): parity with domain experts Type model (10.x): 67% fix rate
English
0
0
0
25
H2LooP
H2LooP@h2loopai·
The real test: fix violations without rewriting the codebase. Character delta on fixes: Sanitizr: ~12% Gemini 2.5 Flash: 25-31% Surgical edits. Near-identical to expert corrections.
English
0
0
0
16
H2LooP
H2LooP@h2loopai·
Common assumption: full fine-tuning performs better, LoRA is the efficient trade-off. We tested this. The assumption is wrong.
H2LooP tweet media
English
4
1
3
36
H2LooP
H2LooP@h2loopai·
The more relevant weight matrices you apply LoRA to, the better the result. The instinct to restrict LoRA to attention layers is leaving performance on the table.
English
0
0
0
13
H2LooP
H2LooP@h2loopai·
With proper hyperparameter setup, LoRA outperforms full fine-tuning while keeping its compute advantages. What "proper" means: Learning rates 10x higher than full SFT Higher rank selection LoRA on all weight matrices, not just attention and proper warmup scheduling + weight decay
English
0
0
0
15
H2LooP
H2LooP@h2loopai·
None of it was obvious before the runs.
English
0
0
0
13
H2LooP
H2LooP@h2loopai·
High-rank LoRA beats low-rank. Domain-only data beats mixed corpora. Full-module targeting is optimal.
English
0
0
0
12