Viex

3 posts

Viex banner
Viex

Viex

@viexonapp

Institutional stablecoin treasury infrastructure. No Coin

🟢 Katılım Nisan 2026
11 Takip Edilen44 Takipçiler
Sabitlenmiş Tweet
Viex
Viex@viexonapp·
🪙 Viex introduces a compliant stablecoin treasury built on Solana, where KYC, AML, and Travel Rule checks are enforced directly in the transfer layer. Every transaction is evaluated before settlement enabling institutions to operate cross border stablecoin liquidity within verifiable compliance boundaries.
English
23
3
29
491
Viex
Viex@viexonapp·
⏳ Everyone focuses on model capability, but the real bottlenecks in production are scheduling, latency, and system coordination. This kind of infra thinking will matter a lot as AI systems start running real workflows. @toly 👏
toly 🇺🇸@toly

MetaTimer: Using Large Language Models for Precise, Prompt-Aware Inference Latency Prediction The rapid proliferation of large language models (LLMs) in production systems has exposed a fundamental limitation: inference latency varies dramatically across prompts due to differences in semantic complexity, required reasoning depth, output length, and generation dynamics. Conventional prediction methods—ranging from token-count heuristics and hardware Roofline models to traditional machine-learning regressors—fail to generalize because they cannot capture these prompt-specific nuances. Accurate a priori estimation of processing time is essential for resource scheduling, dynamic batching, cost forecasting, service-level guarantees, and user-experience enhancements. We introduce MetaTimer, the first framework to repurpose a lightweight LLM itself as a high-precision meta-predictor capable of forecasting the exact wall-clock inference duration required by any target LLM for an arbitrary input prompt. A compact 8B-parameter model is fine-tuned on a massive corpus of millions of prompt–execution pairs collected across heterogeneous model families (GPT-4-class, Llama 3.1, Claude, Mistral), quantization levels, decoding strategies, and hardware accelerators. The predictor employs chain-of-thought reasoning to decompose prompt semantics, estimate output token distributions and reasoning trajectories, and integrate model- and hardware-specific performance profiles, yielding fine-grained predictions for Time-to-First-Token (TTFT), Time-Per-Output-Token (TPOT), and total latency. Extensive evaluations on held-out benchmarks spanning reasoning, creative writing, coding, and long-context tasks demonstrate state-of-the-art accuracy: a mean absolute percentage error (MAPE) of 6.3% for end-to-end latency—representing a >40% reduction in mean squared error relative to the strongest Roofline–ML baselines—and strong zero-shot generalization to unseen models and platforms. When integrated into production serving stacks (vLLM, TensorRT-LLM, Triton), MetaTimer delivers up to 31% gains in resource utilization and tail-latency reduction. These results establish that LLMs possess emergent capabilities for computational self-modeling, opening a new paradigm for self-aware, adaptive, and energy-efficient generative AI infrastructure. We publicly release the predictor model, dataset, and serving plugins to accelerate research in meta-performance modeling for frontier AI systems.

English
12
2
13
195
Viex
Viex@viexonapp·
💻 This is the control layer behind institutional stablecoin treasury on Solana. x.com/Fried_rice/sta… const allowed = senderKyc.active && receiverKyc.active && senderKyc.expiresAt > now && receiverKyc.expiresAt > now && !amlBlacklist.has(from) && !amlBlacklist.has(to) Built for governed cross border settlement
Chaofan Shou@Fried_rice

Claude code source code has been leaked via a map file in their npm registry! Code: …a8527898604c1bbb12468b1581d95e.r2.dev/src.zip

English
11
2
17
296