Aleph Alpha

485 posts

Aleph Alpha banner
Aleph Alpha

Aleph Alpha

@Aleph__Alpha

Our mission is a European generalizable AI. We're hiring: https://t.co/k7MxJK1XU1 #AGI, #artificialintelligence, #writtenbyahuman,#writtenbyanAI

Heidelberg, Germany Katılım Aralık 2018
1 Takip Edilen8.6K Takipçiler
Aleph Alpha
Aleph Alpha@Aleph__Alpha·
Introducing Alpha-MoE: A fused megakernel for faster tensor parallel inference. With up to 200% speedups for MoE models in TP deployments. Optimized for Hopper. Built for sovereignty and scale. aleph-alpha.com/alpha-moe-a-me…
Aleph Alpha tweet media
English
3
18
124
8.4K
Aleph Alpha
Aleph Alpha@Aleph__Alpha·
Inference isn’t just about speed. Our latest blog breaks down how hardware choices impact latency, throughput & cost when serving massive models like DeepSeek v3. Read the blog & full report: aleph-alpha.com/deepseek-infer…
Aleph Alpha tweet media
English
2
2
31
5.2K
Aleph Alpha
Aleph Alpha@Aleph__Alpha·
We’re shipping both models with: Hugging Face inference for easy testing vLLM fork optimized for HAT – note this is still under active development Weights & inference code are available today under Open Aleph license for research and educational use.
English
1
0
21
3.8K
Aleph Alpha
Aleph Alpha@Aleph__Alpha·
Introducing two new tokenizer-free LLM checkpoints from our research lab: TFree-HAT 7B Built on our Hierarchical Autoregressive Transformer (HAT) architecture, these models achieve top-tier German and English performance while processing text on a UTF-8 byte level.
Aleph Alpha tweet media
English
14
46
435
74.2K
Aleph Alpha
Aleph Alpha@Aleph__Alpha·
We signed it, sending a signal to the world. The EU’s AI Code of Practice isn’t just another piece of regulation. It’s a declaration. Of what AI can be when built with transparency, responsibility and sovereignty. Just like ours. aleph-alpha.com/a-signature-an…
English
2
3
12
2.1K
Aleph Alpha
Aleph Alpha@Aleph__Alpha·
For #ICLR2025 we are unveiling a new, high-quality pretraining dataset for German LLMs. Shared to strengthen the open research community. Shaped by our belief in excellence and transparency. huggingface.co/datasets/Aleph…
Aleph Alpha tweet media
English
3
8
45
6.1K
Aleph Alpha
Aleph Alpha@Aleph__Alpha·
@jphme @kerstingAIML Thanks for your interest in our technology! It’s open for research and education to promote transparency and progress in the field. While commercial use is restricted, we fully support innovation and are always open to collaboration and dialogue. Let’s connect!
English
0
0
4
93
Jan P. Harries
Jan P. Harries@jphme·
@Aleph__Alpha @kerstingAIML The "Open Aleph" license states you may use this (scraped) web dataset for "for any Non-Commercial and Non-Administrative purpose." Pretty Lame (and misleading license name) tbh!
English
3
0
7
348
Aleph Alpha
Aleph Alpha@Aleph__Alpha·
🐍 #PyCon has kicked off! Swing by our booth to chat about your career aspirations and explore how we can collaborate 🤝 Let us know you're coming by filling out this form: forms.office.com/Pages/DesignPa…
Aleph Alpha tweet media
English
0
0
4
1.3K
Aleph Alpha
Aleph Alpha@Aleph__Alpha·
🚀 Exciting Announcement from Davos: Aleph Alpha Unveils Tokenizer-Free LLMs! 🚀 We’re thrilled to announce a pioneering innovation that was unveiled yesterday at the World Economic Forum in Davos: Aleph Alpha has introduced a groundbreaking tokenizer-free (T-Free) LLM architecture, setting a new benchmark for sovereign AI solutions. 🔑 What makes this breakthrough so important? Many governments and enterprises have struggled to develop and master a sovereign generative AI strategy due to a lack of data, resources and expertise, forcing them to rely on established models that do not perform adequately on relevant languages and specialized knowledge. This results in a loss of power where value capture, but also ideological alignment, has to be left to a small group of LLM producers. By removing the tokenizer – a bottleneck to adaptability in traditional models – our new tokenizer-free AI architecture enables: ✅ High-quality multilingual capabilities for multiple languages and alphabets. ✅ Customization for specialized industries. ✅ Lower computational demands for fine-tuning, reducing costs and environmental impact. For Finnish, for example, this innovative AI architecture can deliver a 70 percent reduction in total training cost and carbon footprint compared to alternative options. This innovation democratizes access to AI, empowering governments and businesses to create solutions that are fully aligned with their cultural and linguistic specifications, specialized knowledge and terminology, while increasing efficiency and fairness in AI applications. 🌍 💡 What drove us to this development? As a European company rooted in a diverse linguistic and cultural ecosystem, we’ve prioritized sovereignty and multilingualism since day one. Years of research have culminated in the T-Free architecture, a true milestone in AI technology. 🤝 Collaborating with @AMD & Schwarz Digits In collaboration with AMD and Schwarz Digits, Aleph Alpha's new and innovative LLM architecture is enhanced with high performance computing and a sovereign cloud solution, creating a full-stack sovereign offering that combines AI software and hardware. 🔗 Press release: lnkd.in/eGAsPk9R 👉 Technical insights on our blog: lnkd.in/evse6YD9 👉 Our latest peer-reviewed Research Paper: lnkd.in/ewmJTQTf Let’s shape the future of trustworthy, sovereign AI solutions together.
Aleph Alpha tweet mediaAleph Alpha tweet media
English
5
14
80
34.6K
Aleph Alpha
Aleph Alpha@Aleph__Alpha·
New in Neural Network Parametrization Technique: Introducing Unit-Scaled Maximal Update Parametrization (u-μP). In partnership with @GCResearchTeam, u-μP merges μP and Unit Scaling to boost training stability & hyperparameter transfer across model sizes. Read more about the theory & practical tips for LLM training in our blog. 🔗aleph-alpha.com/in-awe-at-the-… #writtenbyalephalpha
Aleph Alpha tweet media
English
2
12
59
9K
Aleph Alpha
Aleph Alpha@Aleph__Alpha·
Our innovation, T-Free, offers a novel approach to tokenization, boosting tokenizer fertility across various languages, and reducing the size of the embedding layer by up to 75% compared to traditional tokenizers. Early experiments with T-Free show promising results and could unlock new possibilities in LLMs, including: - Up to 50% reduction in training and inference costs - Improved semantic encoding of language - Enhanced performance in multilingual models (2/3)
English
1
2
19
3.8K
Aleph Alpha
Aleph Alpha@Aleph__Alpha·
Today we introduce T-Free, a new paradigm in language processing. Tokenization is one of the core building blocks of large language models (LLMs), transforming natural language into numeric representations for further processing. (1/3) 🔗 lnkd.in/eTi7kjuc #writtenbyalephalpha
English
1
35
142
18.1K