Roee Hendel retweetledi
Roee Hendel
12 posts

Roee Hendel retweetledi
Roee Hendel retweetledi
Roee Hendel retweetledi

We released the #Jamba 1.5 open model family:
- 256K #contextwindow
- Up to 2.5X faster on #longcontext in its size class
- Native support for structured JSON output, function calling, digesting doc objects & generating citations
twtr.to/giIEE
#AI #LLM #AI21Jamba


English
Roee Hendel retweetledi

Introducing Jamba, our groundbreaking SSM-Transformer open model!
As the first production-grade model based on Mamba architecture, Jamba achieves an unprecedented 3X throughput and fits 140K context on a single GPU.
🥂Meet Jamba ai21.com/jamba
🔨Build on @huggingface

English
Roee Hendel retweetledi

@yalishandi @_akhaliq We haven't explored this in depth, but it's a promising direction. Studies linking ICL to SGD could offer clues. In related tests, ICL sometimes acts like an empirical risk minimizer, fitting random examples, while other times ignoring them. Further exploration is needed.
English
Roee Hendel retweetledi

In-Context Learning Creates Task Vectors
paper page: huggingface.co/papers/2310.15…
In-context learning (ICL) in Large Language Models (LLMs) has emerged as a powerful new learning paradigm. However, its underlying mechanism is still not well understood. In particular, it is challenging to map it to the "standard" machine learning framework, where one uses a training set S to find a best-fitting function f(x) in some hypothesis class. Here we make progress on this problem by showing that the functions learned by ICL often have a very simple structure: they correspond to the transformer LLM whose only inputs are the query x and a single "task vector" calculated from the training set. Thus, ICL can be seen as compressing S into a single task vector theta(S) and then using this task vector to modulate the transformer to produce the output. We support the above claim via comprehensive experiments across a range of models and tasks.

English

Emergent in-context learning with Transformers is exciting! But what is necessary to make neural nets implement general-purpose in-context learning? 2^14 tasks, a large model + memory, and initial memorization to aid generalization.
Full paper arxiv.org/abs/2212.04458
🧵👇(1/9)

English

@kushal_tirumala Is it possible that the different baseline values simply arise from the fact that larger models have an overall better language modeling capability, rather than memorization? It would be interesting to check the memorization value of the "special batch" prior to training on it.
English


@Carnage4Life The safety of any AI system can be measured by its MtH (meantime to Hitler). Microsoft’s Tay chatbot of several years ago got there in ~24 hours.
English





