Charlie Hou

41 posts

Charlie Hou

Charlie Hou

@hou_char

Research Scientist @ Google DeepMind

เข้าร่วม Eylül 2023
104 กำลังติดตาม45 ผู้ติดตาม
Charlie Hou
Charlie Hou@hou_char·
@_emliu When I was a PhD I also found usually the ROI was good! But sometimes the ROI is not so great.. I once tried applying for a cerebras grant ~2021 and stopped when I realized adapting to their system would be too hard. Though maybe their system is more user friendly now!
English
0
0
0
225
Emmy Liu
Emmy Liu@_emliu·
wrote a guide on getting compute grants as a student, something I wish I did more at the beginning of my PhD. It's honestly one of the highest ROI things you can do as a student (we've gotten 100k+ gpu hrs for roughly 2 weeks of work writing). nightingal3.github.io/blog/2026/04/1…
English
11
139
993
110K
Lucas Beyer (bl16)
Lucas Beyer (bl16)@giffmana·
@tenderizzation I failed a mech eng exam because i brainfarted a comma by one place, and did all the rest of the exam correctly with the wrong number. It was all graded as wrong, because "had you built this, people would've died" Taught me to do sanity checks like a paranoid and not kill people
English
10
5
340
11.8K
Ken Liu
Ken Liu@kenziyuliu·
Sharing a super simple, user-owned memory module we've been playing around: nanomem The basic idea is to treat memory as a pure intelligence problem: ingestion, structuring, and (selective) retrieval are all just LLM calls & agent loops on a on-device markdown file tree. Each file lists a set of facts w/ metadata (timestamp, confidence, source, etc.); no embeddings/RAG/training of any kind. For example: - `nanomem add ` starts an agent loop to walk the tree, read relevant files, and edit. - `nanomem retrieve ` walks the tree and returns a single summary string (possibly assembled from many subtrees) related to the query. What’s nice about this approach is that the memory system is, by construction: 1. partitionable (human/agents can easily separate `hobbies/snowboard.md` from `tax/residency.md` for data minimization + relevance) 2. portable and user-owned (it’s just text files) 3. interpretable (you know exactly what’s written and you can manually edit) 4. forward-compatible (future models can read memory files just the same, and memory quality/speed improves as models get better) 5. modularized (you can optimize ingestion/retrieval/compaction prompts separately) Privacy & utility. I'm most excited about the ability to partition + selectively disclose memory at inference-time. Selective disclosure helps with both privacy (principle of least privilege & “need-to-know”) and utility (as too much context for a query can harm answer quality). Composability. An inference-time memory module means: (1) you can run such a module with confidential inference (LLMs on TEEs) for provable privacy, and (2) you can selectively disclose context over unlinkable inference of remote models (demo below). We built nanomem as part of the Open Anonymity project (openanoymity.ai), but it’s meant to be a standalone module for humans and agents (e.g., you can write a SKILL for using the CLI tool). Still polishing the rough edges! - GitHub (MIT): github.com/OpenAnonymity/… - Blog: openanonymity.ai/blog/nanomem/ - Beta implementation in chat client soon: chat.openanonymity.ai Work done with amazing project co-leads @amelia_kuang @cocozxu @erikchi !!
English
6
36
257
52.7K
Charlie Hou
Charlie Hou@hou_char·
I recently started a new role as a Research Scientist at Google DeepMind! Before looking ahead, I want to say a huge thank you to my former teammates at Meta. I'm incredibly grateful for the time we spent building together, and I know they will continue to build amazing things. I’ve always admired DeepMind’s focus on building AI responsibly to benefit humanity. I'm really excited to collaborate with such a brilliant team and tackle some of the hardest foundational problems in our field. Can't wait to get to work!
Charlie Hou tweet media
English
0
0
4
225
Bojan Jakimovski
Bojan Jakimovski@Shekswess·
Happy to see Federated Learning going big with LLMs !
Charlie Hou@hou_char

Gave a talk at @OpenAI on our work 🌸 POPri “Policy Optimization for Private Data”. POPri is a huge improvement in synthetic data generation under security+privacy constraints! Learn more:

English
1
0
6
325
Matthieu Meeus
Matthieu Meeus@matthieu_meeus·
Pretty clever way to generate DP synthetic text! Previously, private evolution used client feedback to improve synthetic data through in-context learning. This paper uses the client feedback for DPO, as such (i) using the 'negative' samples and (ii) finetuning rather ICL.
Charlie Hou@hou_char

Gave a talk at @OpenAI on our work 🌸 POPri “Policy Optimization for Private Data”. POPri is a huge improvement in synthetic data generation under security+privacy constraints! Learn more:

English
1
0
8
564
Charlie Hou
Charlie Hou@hou_char·
POPri also sees strong efficiency gains vs prior state-of-the-art like Federated Learning, with much lighter computational burden on the user devices. In exchange, more server compute is used.
Charlie Hou tweet media
English
1
0
0
85
Charlie Hou
Charlie Hou@hou_char·
Gave a talk at @OpenAI on our work 🌸 POPri “Policy Optimization for Private Data”. POPri is a huge improvement in synthetic data generation under security+privacy constraints! Learn more:
Charlie Hou tweet media
English
2
2
12
2.3K
Charlie Hou รีทวีตแล้ว
Phillip Wang
Phillip Wang@flippnflops·
I’m excited today to share Grapevine, a system that makes it really easy for AI agents to search company knowledge across, Slack, Notion, codebase, and more. The first app we’re launching with it is a company GPT that works remarkably well, better than any alternative today.
Phillip Wang tweet media
English
14
25
95
52.9K
Charlie Hou รีทวีตแล้ว
Victor Zhou
Victor Zhou@victorczhou·
I've been on the team building Grapevine since the beginning, and in the span of a few months our internal company GPT grew from a hacky experimental demo into my most-used work AI tool.
Phillip Wang@flippnflops

I’m excited today to share Grapevine, a system that makes it really easy for AI agents to search company knowledge across, Slack, Notion, codebase, and more. The first app we’re launching with it is a company GPT that works remarkably well, better than any alternative today.

English
1
3
8
850