Maksim Kuznetsov

29 posts

Maksim Kuznetsov

Maksim Kuznetsov

@Max__Kuznetsov

Research Scientist at @InSilicoMeds

Montréal, Québec Katılım Ocak 2018
145 Takip Edilen125 Takipçiler
Maksim Kuznetsov retweetledi
Sarath Chandar
Sarath Chandar@apsarathchandar·
I am excited to share that our BindGPT paper won the best poster award at @RealAAAI #AAAI2025! Congratulations to the team! Work led by @artemZholus!
Sarath Chandar tweet media
Sarath Chandar@apsarathchandar

What's the foundational model for generative chemistry? Our work, BindGPT, is a good candidate, and it will be presented at #AAAI2025 today! We built a simple transformer language model that beats diffusion models by just generating 3D molecules as text! Led by @artemZholus 1/n

English
3
4
66
5.6K
Maksim Kuznetsov
Maksim Kuznetsov@Max__Kuznetsov·
7/ Finally, nach0-pc enables de novo ligand generation, designing molecules that bind to protein pockets.
Maksim Kuznetsov tweet media
English
0
0
2
113
Maksim Kuznetsov
Maksim Kuznetsov@Max__Kuznetsov·
6/ By injecting noise into point clouds, nach0-pc can generate alternative molecular structures that retain reference molecule shape.
Maksim Kuznetsov tweet media
English
1
0
2
114
Maksim Kuznetsov
Maksim Kuznetsov@Max__Kuznetsov·
1/ At @InSilicoMeds, we’re exploring how language models can process and generate 3D molecular structures. nach0-pc fuses a specialized text-based representation with a domain-specific encoder, enabling precise generation and conditioning on 3D molecular structures.
Maksim Kuznetsov tweet media
English
1
2
6
403
Biology+AI Daily
Biology+AI Daily@BiologyAIDaily·
Bio2Token: All-atom tokenization of any biomolecular structure with Mamba @FlagshipPioneer • This paper introduces “Bio2Token”, a method that tokenizes biomolecular structures at an all-atom level using Mamba. Unlike many current approaches that rely on coarse-grained residue-level representations, Bio2Token focuses on a more detailed atomic-level tokenization. • The innovation here lies in the use of quantized auto-encoders that learn atom-level representations, achieving reconstruction accuracies below and around 1 Ångström. • Mamba, a state space model, plays a key role by providing efficient and scalable encoding, overcoming computational limitations of traditional transformer-based models. Bio2Token can handle structures up to 95,000 atoms, which is significantly larger than the limit for many transformer models. • This approach not only achieves high accuracy but also uses fewer parameters and training resources compared to existing methods like AlphaFold-3 and ESM-3. • Bio2Token demonstrates versatility by tokenizing proteins, RNA, and small molecules, making it a flexible tool for biomolecular structure representation. • The quantized auto-encoders (QAE) efficiently transform 3D structures into 1D discrete tokens, allowing future integration with language models for biomolecular tasks. • The authors present domain-specific tokenizers (mol2token, protein2token, RNA2token) and a combined tokenizer (bio2token) that generalizes across different types of biomolecules. • Compared to ESM-3, Bio2Token achieves a lower reconstruction RMSE and superior performance across protein and RNA datasets, demonstrating its potential as a robust tool for accurate structural modeling. • The combination of Mamba-based architecture and quantized auto-encoders provides a lightweight yet powerful solution, avoiding the quadratic computational cost seen in transformers. • Limitations include ensuring chemical validity in reconstructed structures, as even small deviations can lead to unrealistic bonding. Future directions involve improving accuracy by adding more training data and integrating post-processing steps for chemical validity. @oliviaviessmann 📜Paper: arxiv.org/abs/2410.19110 #biomoleculardesign #proteinmodeling #machinelearning #stateSpaceModel #bioinformatics #Mamba #tokenization
Biology+AI Daily tweet media
English
1
40
188
43.2K
Maksim Kuznetsov
Maksim Kuznetsov@Max__Kuznetsov·
Happy to present our latest results in @InSilicoMeds on molecular graph generation at #AAAI2021! Check out our joint work with @d_polykovskiy “MolGrow: A Graph Normalizing Flow for Hierarchical Molecular Generation” at poster session on 5-Feb, 08:45-10:30 AM & 04:45-06:30 PM PST
Maksim Kuznetsov tweet media
English
0
3
19
0
Maksim Kuznetsov
Maksim Kuznetsov@Max__Kuznetsov·
@norpadon Для сложных пайплайнов есть ещё dvc.org, но его удобность под вопросом
Русский
1
0
0
0
Artur Chakhvadze
Artur Chakhvadze@norpadon·
@Max__Kuznetsov У меня слишком сложные пайплайны, там лайтнинга недостаточно
Русский
1
0
0
0
Artur Chakhvadze
Artur Chakhvadze@norpadon·
Каждый раз когда я начинаю новый ML проект я по нескольку дней аутирую потому что не знаю как написать трейнинг луп. Как сделать построение модели из конфига, как тестировать, как оранизовать пайплайн чтобы его не пришлось переписывать если захочется воткнуть в середину ган.
London, England 🇬🇧 Русский
1
0
8
0
Maksim Kuznetsov
Maksim Kuznetsov@Max__Kuznetsov·
@norpadon Чтобы избавиться от проблем с трейн лупом, наши предки принимали простой советский...
Русский
1
0
1
0