Filip Graliński

204 posts

Filip Graliński

Filip Graliński

@FilipGralinski

6502 and Haskell hacker, machine learner, hypopolyglot (many languages, all poor), opposite Pole, skeptical forteanist

Katılım Eylül 2021
587 Takip Edilen111 Takipçiler
Filip Graliński retweetledi
Snowflake
Snowflake@Snowflake·
Day 2 of #SnowflakeSummit flew by but not before a mountain of announcements from our Platform Keynote! We announced: Adaptive Compute, Snowflake Openflow, Cortex AISQL, Semantic Model Sharing, Snowflake Intelligence, and much more. See what's new: bit.ly/4mNjiqR
English
3
3
12
2K
Filip Graliński
Filip Graliński@FilipGralinski·
@jxmnop Yeah, great paper, also for educational purposes. I'm curious whether the embeddings are still available... Whether king + woman - man = queen was already there but they didn't realize it...
English
0
0
0
206
dr. jack morris
dr. jack morris@jxmnop·
did you know people have been training neural networks on text since 2003? everyone talks about Attention Is All You Need. but this is the real paper that got our field started. it was in 2003, in montreal. i read it, and it was even more forward-thinking than i expected:
dr. jack morris tweet media
English
34
111
1.3K
94.9K
Filip Graliński
Filip Graliński@FilipGralinski·
@spacemanidol Todo lenguaje es un alfabeto de símbolos cuyo ejercicio presupone un pasado que los interlocutores comparten...
Español
0
0
0
22
Daniel Campos
Daniel Campos@spacemanidol·
Kinda crazy how much random literature applies to the LLM world now. Borges would probably have a field day with LLMs and the constant talk of the incoming singularity. Truly the creation of library of babel.
English
1
0
1
262
Filip Graliński retweetledi
Łukasz Borchmann
Łukasz Borchmann@LukaszBorchmann·
How can the most accurate SQL be generated for a given question? We propose a method to significantly boost text-to-SQL accuracy while drastically cutting costs.👇 #NLProc #AI #TextToSQL #LLMs
Łukasz Borchmann tweet media
English
2
3
54
51.2K
Filip Graliński retweetledi
Anupam Datta
Anupam Datta@datta_cs·
Our Snowflake AI Research team just released Arctic Embed’s core training code into the open source ArcticTraining project — making it easier for developers and researchers to reproduce, fine-tune, and build on our embedding models. Arctic Embed is the leading small embedding model on the MTEB leaderboard and is widely used with over 1M monthly downloads. What you’ll find: ✅ Clean, config-driven workflows powered by DeepSpeed ✅ Flexible contrastive data handling ✅ Example fine-tuning recipes and ready-to-use tooling Read more here and try it out: snowflake.com/en/engineering… @SnowflakeDB @DeepSpeedAI @lukemerrick_ @pxyumass @spacemanidol @rajhans_samdani @jeffra45 @StasBekman
English
4
5
15
10.1K
Filip Graliński retweetledi
Luke Merrick
Luke Merrick@lukemerrick_·
Connor Shorten was kind enough to give me the mic for a lot of hot takes on text embedding models in the latest Weaviate podcast.
Connor Shorten@CShorten30

Arctic Embed ❄️ has been one of the most impactful open-source text embedding models! In addition to the open model, which has helped a lot of companies kick off their own inference and fine-tuning services (including us), the Snowflake team has also published incredible research breaking down all the components of how to train these models! I am SUPER EXCITED to publish the 110th Weaviate Podcast with Luke Merrick (@lukemerrick_), Puxuan Yu (@pxyumass), and Charles Pierse (@cdpierse) discussing all things Arctic Embed! The podcast covers: • The origin of Arctic Embed • Pre-training embedding models • Matryoshka Representation Learning • Fine-tuning embedding models • Synthetic Query Generation • Hard Negative Mining • Single-Vector Embedding Models in the search model cohort of ColBERT, SPLADE, and Re-rankers I hope you enjoy the podcast! As always, please reach out if you would like to discuss any of these ideas further!

English
1
4
11
1.1K
Chris Henson
Chris Henson@_chenson__·
@gro_tsen Was this a standard notation in the 50s? Seems pretty confusing! (From the first reference)
Chris Henson tweet media
English
2
0
3
864
Gro-Tsen
Gro-Tsen@gro_tsen·
Learned on MathOverflow: it is possible to write a finite formula for n! involving just the operations of addition, subtraction, multiplication, integer division, and exponentiation. Precise statement is here: mathoverflow.net/a/484115/17064
Gro-Tsen tweet media
English
14
64
659
45.2K
Filip Graliński
Filip Graliński@FilipGralinski·
@magoniareview Of interest! Actually, it'd be great if it covered languages other than English... (reading weird books is the best way to learn foreign languages!), I can share a list of recent "Magonian" books in a bunch of languages, if you're interested.
English
0
0
0
19
thePelican
thePelican@magoniareview·
I have now suspended the Pelican's 'Book News' blog, but may resurrect is as a stand-alone listing of forthcoming titles of Fortean, ufological and folkloric interest, if enough people think it will be of interest.
thePelican tweet media
English
3
1
8
648
Filip Graliński retweetledi
Aurick Qiao
Aurick Qiao@aurickq·
We are excited to share SwiftKV, our recent work at @SnowflakeDB AI Research! SwiftKV reduces the pre-fill compute for enterprise LLM inference by up to 2x, resulting in higher serving throughput for input-heavy workloads. 🧵
Aurick Qiao tweet media
English
4
15
43
2.8K
Kamil Galeev
Kamil Galeev@kamilkazani·
Obviously Germans had the reputation for copying things (just a bit earlier). Which is a standard, textbook pattern for a rising manufacturing power. Produce cheap, crappy, low quality goods. Copy everything that moves, and doesn’t move. Break every patent and every copyright.
Crémieux@cremieuxrecueil

Pretty insane Community Note. Never mind that the Germans didn't have a reputation for copying like China does, the first two examples here were Jews and the last one was a spy, albeit for the Russians.

English
16
36
372
43.2K
Filip Graliński retweetledi
Daniel Campos
Daniel Campos@spacemanidol·
🚀 I am thrilled to introduce @SnowflakeDB 's Arctic Embed 2.0 embedding models! 2.0 offers high-quality multilingual performance with all the greatness of our prior embedding models (MRL, Apache-2 license, great English retrieval, inference efficiency) snowflake.com/engineering-bl…🌍
English
8
26
80
7.1K
Filip Graliński retweetledi
Michał Pietruszka
Michał Pietruszka@MichaPietruszka·
Can AI models help us create better models? 🧵 1/ It's a question that stands at the boundaries of what's possible in data science. We explored how Large Language Models (LLMs) perform as data scientists, especially in the art of feature engineering.
Michał Pietruszka tweet media
English
1
1
3
192
Filip Graliński
Filip Graliński@FilipGralinski·
@hrishioa @Hrishioa clarify what "clarification question" token is in the guide; I was, like, what is that magic token?? (but anyway, cool doc, thanks)
English
1
0
1
225
Hrishi
Hrishi@hrishioa·
Entropix is beyond cool but hard to understand Here's a guide that means you don't have to walk my ragged path: southbridge-research.notion.site/Entropixplaine… Spent an hour talking to imaginary AI friends about the code, then ran it through Lumentis
English
21
47
475
81.1K
Filip Graliński
Filip Graliński@FilipGralinski·
@hrishioa I want to know... what the token 2564 is... ... I'd prefer 1981, for esthetic reasons
English
0
0
1
106