Patrick Bunk retweetledi
Patrick Bunk
896 posts

Patrick Bunk
@patrickbunk
CEO @Ubermetrics https://t.co/Yffdnq0PNJ
Berlin, Germany Katılım Mayıs 2009
414 Takip Edilen276 Takipçiler

@annewoj23 Cool! @grok please hit me with an overview of this post as if you’re a sports caster and I’m mentally 11.
English

To the 23andMe Community,
I am incredibly excited and humbled to share with you that TTAM Research Institute (TTAM), a new non—profit medical research organization that I founded and lead, has completed the acquisition of 23andMe. I formed TTAM and pursued the acquisition of 23andMe because of my strong belief that everyone should continue to have the ability to learn about and benefit from their DNA, and that a non-profit structure is the best way to solidify our values and commitments to our customers, the scientific community and the world at large.
Since its founding in 2006, 23andMe has stood for something revolutionary: that individuals should have a right to understand themselves at the most fundamental level — through their DNA. Through their DNA they can understand their health risks , prevent disease and be true partners in participating in research that shapes the future of health and medicine.
Understanding the code of life - our DNA - is one of the most exciting scientific missions of our lifetime. Through our DNA, we can learn about our ancestry origins, prevent potential health issues, and help develop treatments that could benefit all of us. Humans are 99.5% genetically identical, and we have DNA in common with everything alive today. The pursuit to understand DNA will benefit all of us and all of life.
That belief built one of the world’s most diverse and engaged genetic research communities. More than 15 million people have joined this journey, and over 80% have chosen to participate in research, contributing to over 275 peer-reviewed publications in cancer, cardiovascular disease, neurological disorders, and more. These discoveries did not just stay academic - those discoveries were also translated into new genetic reports for customers that helped them continue to learn about their own DNA. Thank you to the 23andMe community that participated in research and contributed to discoveries that have had a meaningful impact for the world.
I want to end on a personal note. Over the last few years I unfortunately lost my father, my nephew and my sister. These tragedies have sharpened my focus on what is most important to me and how I want to spend my time. My passion is 23andMe. The opportunity to give back to society with our research and help everyone benefit from learning about their genome with a healthier life, is a personal mission where I am dedicating my resources and my time. We all have a disease or health condition that we care about . By coming together as a single community, we are stronger and more powerful to make discoveries and to ultimately make a difference. The future of healthcare belongs to all of us and it’s in our power to make a difference.
I am honored to be back with 23andMe and look forward to sharing more with you in the weeks and years to come.
With gratitude,
Anne

English

nobody likes building out an n8n workflow from scratch...
that's why I turned Claude into my personal n8n builder
I just tell it what kind of flow I want (or drop a screenshot) and it:
– builds the entire thing
– adds full documentation
– leaves sticky notes explaining each node
want the Claude prompt + cheat sheets to set this up?
like + comment “claude” and I’ll send it over (must be following so I can DM)
English
Patrick Bunk retweetledi
Patrick Bunk retweetledi
Patrick Bunk retweetledi

@D0R0 Ich habe auf diese Frage in den letzten sechs Jahren noch nie eine Antwort erhalten, so oft ich sie auch gestellt habe.
Deutsch
Patrick Bunk retweetledi

Announcing reader-lm-0.5b and reader-lm-1.5b, jina.ai/news/reader-lm… two Small Language Models (SLMs) inspired by Jina Reader, and specifically trained to generate clean markdown directly from noisy raw HTML. Both models are multilingual and support a context length of up to 256K tokens. Despite their compact size, these models achieve state-of-the-art performance on this HTML2Markdown task, outperforming larger LLM counterparts while being only 1/50th of their size.
English
Patrick Bunk retweetledi
Patrick Bunk retweetledi

The reason I’m insanely bullish on AI is that since starting Box, we have never seen a bigger shift in how we can work with our enterprise information than today.
AI completely revolutionizes how we can work with enterprise information. Since the mainframe era, it’s been relatively trivial to work with our *structured* data in an enterprise. We could query, compute, synthesize, summarize, and analyze anything that could be structured in a database - i.e. the data sitting in our ERP, CRM, and HR systems.
But it turns out this is only a small fraction of our corporate information. If you were to “weigh” the amount of data inside of an enterprise (in the form of raw storage), roughly 10% of it would be structured data, and 90% of it would be unstructured data. And our content — things like our documents, contracts, product specs, financial records, marketing assets and videos — makes up the vast majority of this corporate data. Yet for essentially the entire history of computing, we haven’t *really* been able to make sense of this information unless a human is involved. Of course we can store it, send it, share it, and search for it — but deeply understanding what’s inside this information in a way that computers can interact with intelligently has been near-impossible.
Well, for the first time ever, generative AI actually lets us talk to our unstructured data. Multimodal models especially allow us to process this content using a computer and essentially perform any task that a human can, but at infinite scale and speed. This is utterly game-changing when working with information in the enterprise.
Instantly, our content goes from being digital artifacts that get touched once in a while, to a digital memory that anyone in the enterprise can tap into always. All of a sudden instead of the more information you have making things harder to find and make sense of, the opposite becomes true. And we enter a world where your digital information becomes one of your most valuable resources.
When we can turn our content into valuable knowledge, everything about how we work changes. A new employee instantly has access to the same expertise of someone who’s worked at a company for 15 years; when you can understand what’s inside of content — like contracts, invoices, or digital assets— and extract its structured data, you can automate nearly any workflow; and AI can let us classify and protect content with a level of precision that’s never been possible before to prevent threats and risks across the enterprise.
This is simply the biggest change we’ve ever seen with how we can work with our data, and this is what we’re building with Box AI.
In 1945, Vannevar Bush wrote a seminal article which outlined eerily insightful predictions, including the idea of the “Memex”, a new device “in which an individual stores all his books, records, and communications, and which is mechanized so that it may be consulted with exceeding speed and flexibility. It is an enlarged intimate supplement to his memory.”
The vision laid out imagined a future where the more knowledge and information your “computer” had, the smarter and more informed you would become. While many aspects of PCs, mobile devices, and the cloud eventually resembled this early vision, the seamlessness in how we could work with our information never quite played out.
Until today.
English
Patrick Bunk retweetledi

# CUDA/C++ origins of Deep Learning
Fun fact many people might have heard about the ImageNet / AlexNet moment of 2012, and the deep learning revolution it started.
en.wikipedia.org/wiki/AlexNet
What's maybe a bit less known is that the code backing this winning submission to the contest was written from scratch, manually in CUDA/C++ by Alex Krizhevsky. The repo was called cuda-convnet and it was here on Google Code:
code.google.com/archive/p/cuda…
I think Google Code was shut down (?), but I found some forks of it on GitHub now, e.g.:
github.com/ulrichstern/cu…
This was among the first high-profile applications of CUDA for Deep Learning, and it is the scale that doing so afforded that allowed this network to get such a strong performance in the ImageNet benchmark. Actually this was a fairly sophisticated multi-GPU application too, and e.g. included model-parallelism, where the two parallel convolution streams were split across two GPUs.
You have to also appreciate that at this time in 2012 (~12 years ago), the majority of deep learning was done in Matlab, on CPU, in toy settings, iterating on all kinds of learning algorithms, architectures and optimization ideas. So it was quite novel and unexpected to see Alex, Ilya and Geoff say: forget all the algorithms work, just take a fairly standard ConvNet, make it very big, train it on a big dataset (ImageNet), and just implement the whole thing in CUDA/C++. And it's in this way that deep learning as a field got a big spark. I recall reading through cuda-convnet around that time like... what is this :S
Now of course, there were already hints of a shift in direction towards scaling, e.g. Matlab had its initial support for GPUs, and much of the work in Andrew Ng's lab at Stanford around this time (where I rotated as a 1st year PhD student) was moving in the direction of GPUs for deep learning at scale, among a number of parallel efforts.
But I just thought it was amusing, while writing all this C/C++ code and CUDA kernels, that it feels a bit like coming back around to that moment, to something that looks a bit like cuda-convnet.

English
Patrick Bunk retweetledi

@Culture_Crit I am experiencing physical pain reading through this
English
Patrick Bunk retweetledi
Patrick Bunk retweetledi

Court filing in #OpenAI case reflects NYT’s fundamental problem with #copyright law: facts are free for the taking
NYT opposition to OpenAI motion for partial dismissal mischaracterizes ChatGPT and ignores that U.S. copyright law has clear limits.
aifray.com/court-filing-i…

English
Patrick Bunk retweetledi

# On the "hallucination problem"
I always struggle a bit with I'm asked about the "hallucination problem" in LLMs. Because, in some sense, hallucination is all LLMs do. They are dream machines.
We direct their dreams with prompts. The prompts start the dream, and based on the LLM's hazy recollection of its training documents, most of the time the result goes someplace useful.
It's only when the dreams go into deemed factually incorrect territory that we label it a "hallucination". It looks like a bug, but it's just the LLM doing what it always does.
At the other end of the extreme consider a search engine. It takes the prompt and just returns one of the most similar "training documents" it has in its database, verbatim. You could say that this search engine has a "creativity problem" - it will never respond with something new. An LLM is 100% dreaming and has the hallucination problem. A search engine is 0% dreaming and has the creativity problem.
All that said, I realize that what people *actually* mean is they don't want an LLM Assistant (a product like ChatGPT etc.) to hallucinate. An LLM Assistant is a lot more complex system than just the LLM itself, even if one is at the heart of it. There are many ways to mitigate hallcuinations in these systems - using Retrieval Augmented Generation (RAG) to more strongly anchor the dreams in real data through in-context learning is maybe the most common one. Disagreements between multiple samples, reflection, verification chains. Decoding uncertainty from activations. Tool use. All an active and very interesting areas of research.
TLDR I know I'm being super pedantic but the LLM has no "hallucination problem". Hallucination is not a bug, it is LLM's greatest feature. The LLM Assistant has a hallucination problem, and we should fix it.
Okay I feel much better now :)
English
Patrick Bunk retweetledi
Patrick Bunk retweetledi
Patrick Bunk retweetledi











