Günter Klambauer Profile
Günter Klambauer

@gklambauer

2,525
Followers
409
Following
95
Media
1,461
Statuses

Deep Learning researcher known for self-normalizing neural networks, applications of Machine Learning in Life Sciences areas; ELLIS program Director

Joined December 2021
Don't wanna be here? Send us removal request.
Pinned Tweet
@gklambauer
Günter Klambauer
25 days
VN-EGNN: E(3)-Equivariant GNNs with Virtual Nodes Enhance Protein Binding Site Identification New method to find binding pockets of proteins. Virtual nodes allow to employ distance losses directly. P: C: 🤗:
6
50
169
@gklambauer
Günter Klambauer
5 months
HITCHHIKER'S GUIDE Comprehensive overview of Geometric Deep Learning approaches to molecular system! Paper:
Tweet media one
0
48
258
@gklambauer
Günter Klambauer
2 years
The ELLIS ML4Molecules workshop will also happen this year on November 28 in VIRTUAL format! Please find the announcement and the call for papers here: Looking forward to your contributions!
Tweet media one
2
86
234
@gklambauer
Günter Klambauer
1 year
Diffusion models used to generate realistically looking microscopy images of cells:
Tweet media one
4
31
177
@gklambauer
Günter Klambauer
8 months
🎾🥳DEEP LEARNING GRAND SLAM🥳🎾 (get a paper accepted at ICLR, ICML, NeurIPS within one year 😜) ICLR2023: [MHNfs]() ICML2023: [CLAMP]() NeurIPS2023: Initialization for Input-Convex Nets Credits should go to the PhD students 🎓👏
Tweet media one
1
14
170
@gklambauer
Günter Klambauer
1 year
Large language models can generate new proteins -- even quite different from all known ones -- that show catalytic activity in wet-lab tests! Wow!
1
27
164
@gklambauer
Günter Klambauer
1 year
Finally, someone made the decision-tree learning differentiable. Reformulation of the classification function to dense representations; approximation of step function with sigmoids, entmax function etc. Good results on a large number of datasets.
2
32
160
@gklambauer
Günter Klambauer
8 months
Generating molecule with a given 3D shape. The method ShapeMol uses a conditional diffusion probabilistic model and equivariant layers. Benchmarked on experiments suggested by SQUID.
Tweet media one
0
29
144
@gklambauer
Günter Klambauer
5 months
### MACHINE LEARNING FOR MOLECULES ### This Friday (Dec 8), at 9am in European Central Time zone!! Free to join for everyone! Program available here:
0
40
125
@gklambauer
Günter Klambauer
2 years
Self-supervised pre-trained networks are usually more robust to distribution shifts than networks pre-trained in supervised or un-supervised fashion. Thorough analysis of this behaviour:
1
19
121
@gklambauer
Günter Klambauer
4 months
Multi-Modal Representation Learning for Molecular Property Prediction: Sequence, Graph, Geometry Three chemical modalities are contrasted against each other and used for property prediction. Unfortunately, only evaluated on the MoleculeNet benchmarks
Tweet media one
0
24
120
@gklambauer
Günter Klambauer
7 months
Bayesian Deep Learning (BDL) Library released: Use BDL easily: e.g. MC-Dropout with a vision transformer can readily be coded in few lines Paper: Repo:
Tweet media one
0
20
115
@gklambauer
Günter Klambauer
8 months
New physics-inspired descriptors for non-bonded interactions. Overcoming limitations of local geometry-based ML approaches, this method incorporates long-range effects with a focus on diverse non-bonded potentials. 🚀📊
Tweet media one
0
23
111
@gklambauer
Günter Klambauer
10 months
Honored to be invited to talk about ArtificiaI Intelligence in the Austrian parliament! Will provide a re-recording of my talk here soon. Should it be English or German?
Tweet media one
5
6
107
@gklambauer
Günter Klambauer
2 years
Open positions (PostDoc/PhD) in machine learning and deep learning at the LIT AI Lab and the ELLIS unit Linz: Please share!
2
57
100
@gklambauer
Günter Klambauer
9 months
PREFER, a Python-based framework powered by AutoSklearn, assists molecular property prediction. Effortlessly compare diverse representations and ML models for accelerated discovery. Open-source on GitHub. 🧪🔬 #Cheminformatics #ML Details:
1
26
101
@gklambauer
Günter Klambauer
1 year
At NeurIPS, @HochreiterSepp critized LLMs for using the parameters for storing phrases; there would be better ways, such as a modern Hopfield network, to store this. Well, here is a retrieval system with 500x less parameters that follows this approach:
2
16
96
@gklambauer
Günter Klambauer
1 year
Workshop program & registration (FREE): See you in a week!
Tweet media one
6
45
95
@gklambauer
Günter Klambauer
9 months
Our method CLAMP (Contrastive Language Assay Molecule Pre-training, ICML2023) uses the same loss:
@giffmana
Lucas Beyer (bl16)
9 months
What makes CLIP work? The contrast with negatives via softmax? The more negatives, the better -> large batch-size? We'll answer "no" to both in our ICCV oral🤓 By introducing SigLIP, a simpler CLIP that also works better and is more scalable, we can study the extremes. Hop in🧶
Tweet media one
27
290
2K
0
16
98
@gklambauer
Günter Klambauer
1 year
Geometric Learning concepts (Equivariance, Invariances) carried over to interpretability-methods: if a neural net is invariant to a certain transformation of the input, also the feature-importance should be invariant. Overly complicated notation, imho
1
17
96
@gklambauer
Günter Klambauer
1 year
I really appreciate this one:
0
17
92
@gklambauer
Günter Klambauer
1 year
AlphaFold is also vulnerable to adversarial attacks: small changes in the protein sequence can lead to drastic changes in folding structure. Shown on data with controlled changes to the protein sequence and on COVID-19 protein variants:
1
16
88
@gklambauer
Günter Klambauer
2 years
One GNN encodes the graph in the usual way; a second GNN gets the input graph with an adjacency matrix, in which edges are determined by node similarity (kNN). Contrastive learning forces the GNN to preserve the structure:
2
19
82
@gklambauer
Günter Klambauer
1 year
This work reports strong improvements at "protein design": This task is defined as predicting amino-acid sequence from 3D structure (coordinates) -- that means it's a kind of *inverse-AlphaFold*.
1
19
84
@gklambauer
Günter Klambauer
8 months
The 3rd edition of the ELLIS ML4Molecules Workshop has been announced! ⏰ Virtual event on December 8 ⏰ Call for contributions open! Participation is free -- join us!!!
Tweet media one
0
29
83
@gklambauer
Günter Klambauer
5 months
WOW! Diffusion models.. "We redesign the network layers to preserve activation, weight, and update magnitudes on expectation. [..] systematic application of this philosophy [..] results in considerably better networks at equal computational complexity"
Tweet media one
0
14
81
@gklambauer
Günter Klambauer
1 year
Large language model pre-trained on masked SMILES. Then applied to MoleculeNet. Obviously, we can now predict clinical tox with AUC 99.7. Basically all wet-lab tox and clinical trials unnecessary now - LOL. The community must stop to deceive itself..
3
12
79
@gklambauer
Günter Klambauer
1 year
Prediction of "cancer genes" using a Graph Neural Networks on protein-protein-interaction networks, datasets like string-db.
2
16
77
@gklambauer
Günter Klambauer
1 year
Prediction of drug synergies with Deep Learning: now there seem to be sufficient data to predict the synergistic effects for new compounds . Advances the ideas of DeepSynergy ( ).
0
16
75
@gklambauer
Günter Klambauer
1 year
🗜️CLAMP (Contrastive Language-Assay-Molecule Pre-training), a new 🚀 method for *zero-shot drug discovery* that utilizes textual assay descriptions for molecular property prediction. Shows that scientific language models are bad at activity prediction.
Tweet media one
2
26
75
@gklambauer
Günter Klambauer
1 year
New self-supervised learning: instead of predicting the masked part of an image, the method tries to match the embedding of the masked part of the region with the embedding from the context (for the masked region). Sry, difficult without formulas.. ;)
0
11
75
@gklambauer
Günter Klambauer
1 year
A message-passing network (MPNN) with one virtual node, connected to all nodes, can approximate a Transformer layer. Many groups have already used such virtual nodes in the past...
0
14
75
@gklambauer
Günter Klambauer
1 year
We've used CLOOB to develop a search engine that unlocks querying a bioimaging database with chemical structures. The CLOOME encoder produces bioimage embeddings that can clearly distinguish new cell phenotypes: Search engine:
Tweet media one
@fuerst_andreas
Andreas Fürst
1 year
We are excited to announce that our latest work, CLOOB, was accepted at this year's NeurIPS 🎉 CLOOB consistently outperforms CLIP at zero-shot transfer learning on a large variety of datasets. 🧵 1/4
Tweet media one
2
62
246
1
21
71
@gklambauer
Günter Klambauer
2 years
Can a molecular graph be reconstructed from chemical fingerprints using machine learning? In this study ( ) a Transformer reconstructs molecules (SMILES) from ECFP4. Only 1-4% of SMILES could be correctly reconstructed -- this should work much better...
11
12
69
@gklambauer
Günter Klambauer
2 years
Open: 5 PhD and 2 postdoc positions in cheminformatics and molecular simulations at University Vienna
0
37
69
@gklambauer
Günter Klambauer
1 year
We are happy that our new ELLIS program "Machine Learning for Molecule Discovery" has been accepted by @ELLISforEurope !
@ELLISforEurope
ELLIS
1 year
Great news! We are excited to announce that our #ELLISforEurope research activities are expanding further! The proposals ‘Machine Learning for Molecule Discovery’ and ‘Learning for Graphics and Vision’ have been accepted as new #ELLISPrograms ! #AI #ML
1
21
109
0
14
68
@gklambauer
Günter Klambauer
2 years
Rich Sutton's new take on AI research:
1
18
68
@gklambauer
Günter Klambauer
6 months
🔬A “Google” for microscopy images and molecules🔬 Given an image of cells treated with a molecule, CLOOME can correctly identify this molecule – a task considered impossible even for human experts! Work by @ana_sanchezf of @AiddOne . Paper:
Tweet media one
1
13
69
@gklambauer
Günter Klambauer
8 months
Oh, finally someone else also calling for more scientific rigor in the GNN-community :) Methods with and without hyperparameter selection compared on longe-range graph benchmark. In 2018, we also described this "hyperparameter selection bias"..
2
18
68
@gklambauer
Günter Klambauer
7 months
In-Context Learning for Drug Discovery: Embedding-based few-shot learning methods are equivalent to "in-context learning" of LLMs. Here this concept is used again (but introduced before in MHNfs by Schimunek et al, 2023):
Tweet media one
0
14
68
@gklambauer
Günter Klambauer
7 months
Excited to share our paper “A community effort in SARS-CoV-2 drug discovery" (after 3 years in work)! We report the results of an open science community effort to identify small-molecule inhibitors against SARS-CoV-2. Paper:
Tweet media one
1
13
67
@gklambauer
Günter Klambauer
5 months
LAST REMINDER Tomorrow (Dec 8), 9am CET, this year's @ELLISforEurope Machine Learning for Molecules Workshop starts! Open for everyone to join for free! Schedule and registration:
Tweet media one
5
18
66
@gklambauer
Günter Klambauer
8 months
Predicting molecular activity with XGBoost. 📊🧪 Study on feature importance, highlighting the need for expert interpretation. Hyperparameter optimization is crucial. Valuable guidelines for #cheminformatics practitioners. #ML Paper:
Tweet media one
1
7
63
@gklambauer
Günter Klambauer
7 months
Beam enumeration: an approach for goal-directed molecular generation. Substructures with high likelihood are kept during learning and others are discarded. Language model as basis for the generation process. Paper:
Tweet media one
0
8
63
@gklambauer
Günter Klambauer
1 year
HyperPCM for predicting drug target interactions: Join us today at the poster session of the @AI_for_Science workshop at #NeurIPS2022 . 12:05pm-1pm and 5:10pm-6pm, room388! @AiddOne
Tweet media one
0
11
61
@gklambauer
Günter Klambauer
6 months
Key elements underlying molecular property prediction? Here self-supervised learning methods for small molecules are compared against descriptor-based methods. Another call for scientific rigor and to not over-hype GNNs for small mols..
Tweet media one
2
11
60
@gklambauer
Günter Klambauer
4 months
Pretraining a GNN encoder for molecular structures on a conformation sampling task. Builds on the smart idea (that has been around) to use a diffusion model on the atom coordinates.
1
5
60
@gklambauer
Günter Klambauer
2 years
NVIDIA is also generating small molecules now: ;). VAE-like approach to generate molecules and optimize their properties by sampling from latent space. They are aware of the problems of these optimization cycles ( ).
1
14
58
@gklambauer
Günter Klambauer
1 year
Happy Easter! Luckily, it's Monday and arxiv does not take vacation: A graph neural network approach to predict chemical properties using an initial (bad) 3D conformation. Then gradually improves this conformation with de-noising. Good results on QM9:
1
9
56
@gklambauer
Günter Klambauer
10 months
Study of drug-ranking with anti-cancer properties for particular cell lines characterized by their gene expression profile. Interesting fact 1: bilinear regression is used to combine modalities Fact 2: Morgan FPs plus MLP worked better than GNN (sry folks)
Tweet media one
3
7
56
@gklambauer
Günter Klambauer
1 year
Neural architecture searching (NAS) for new graph neural network architectures. The found architectures (Table 8 in Appendix) are relatively shallow and often LSTMs are selected to aggregate layers. Tests on non-molecule tasks (CiteSeer, PubMed).
0
14
55
@gklambauer
Günter Klambauer
1 year
@andrewwhite01 According to a recent discussion at the ML4Molecules workshop, prediction of binding affinities must be strongly improved...
2
1
53
@gklambauer
Günter Klambauer
5 months
New way of goal-directed optimization of molecular structure: GNNs reformulated as mixed integer linear programming (MILP) problems. Solvers for those problems then provide chemical structures. Paper:
0
9
53
@gklambauer
Günter Klambauer
1 year
OPEN POSITIONS! We have open positions for PhD students and PostDocs at my lab both in core machine learning as well as in "Machine Learning in Life Sciences":
0
34
54
@gklambauer
Günter Klambauer
6 months
🔥🔥🔥 Contrastive learning unleashed 🔥🔥🔥 A powerful, transferable MICROSCOPY IMAGE and MOLECULE ENCODER has been trained on CellPainting data through self-supervised learning (SSL). Paper: Code: App:
2
7
54
@gklambauer
Günter Klambauer
3 months
Thanks for three thousand and one citations! :D
Tweet media one
1
1
52
@gklambauer
Günter Klambauer
7 months
Deep Docking methods now on the rise. This one, too, shows good performance with a relatively standard GNN/MPNN approach. Paper:
Tweet media one
1
5
51
@gklambauer
Günter Klambauer
4 months
HyperPCM: Robust Task-Conditioned Modeling of Drug–Target Interactions Hypernetworks can carry over information from one protein activity prediction task to another. work by @EmmaJMSvensson in @AiddOne
Tweet media one
0
7
51
@gklambauer
Günter Klambauer
1 year
Large dataset for ACTIVITY CLIFF modeling consists of 400K matched molecular pairs. Deep Learning models perform ok, again ECFP plus deep multi-layer perceptrons perform best -- in accordance with works by other groups...
3
9
49
@gklambauer
Günter Klambauer
7 months
Tweet media one
3
4
49
@gklambauer
Günter Klambauer
6 months
Exploring Causality in Single-Cell Genomics 🧬🤖 ML adapts causal techniques to high-dimensional single-cell data, addressing challenges and paving the way for informed experimental design. 📊🔬 Paper:
Tweet media one
0
3
49
@gklambauer
Günter Klambauer
8 months
Multi-modal anything to anything :) NextGPT can take text, image, audio or video as input and generate output in these modalities. Quite efficient because only input projectors and output are trained.
Tweet media one
1
7
48
@gklambauer
Günter Klambauer
1 year
Self-supervised representation learning for time series: a sequence is split into past and future, and the learned representations have to be similar. Does not need negatives -- I don't see why mode collapse should not happen..
0
14
47
@gklambauer
Günter Klambauer
1 year
It's a matter of the initial molecular representations. That's why I always for using ECFP/Morgan plus MLPs as baseline in all studies. Would have saved us from the pile of the zillion molecule encoders that we are now facing...
@KevinKaichuang
Kevin K. Yang 楊凱筌
1 year
Activity cliff: two molecules have similar structure, but a big difference in bioactivity. ML approaches don't do well with activity cliffs, but 'old-school' ML models tend to work better than deep learning. @DerekvTilborg @korney34 @fra_grisoni
Tweet media one
Tweet media two
Tweet media three
Tweet media four
5
29
161
2
1
47
@gklambauer
Günter Klambauer
1 year
🚀 Transforming molecules discovery with few-shot learning 🚀! Our new method **MHNfs** enriches molecule representations with CONTEXT. MHNfs sets a new state-of-the-art on the FS-Mol benchmark dataset. #drugdiscovery #fewshotlearning #AI Paper:
@JSchimunek
Johannes Schimunek
1 year
🚀 Excited to share our #ICLR2023 work on 🚨 context-enriched molecule representations🚦 improve few-shot drug discovery 💊 🚨 Paper: App: HuggingFace 🤗 under prep! #ICLR2023 🧑‍💼 poster 🗨: ⏰ Wed 3 May 4:30 pm - 6:30 pm CAT
Tweet media one
2
23
44
0
11
45
@gklambauer
Günter Klambauer
6 months
Context dependent molecular representations are enabled by LLM-like architectures and self-supervised learning strategies. Performance and benchmarking are still limited. Paper:
Tweet media one
0
8
46
@gklambauer
Günter Klambauer
1 year
I really appreciate this result about the expressiveness of the molecular fingerprint representations. The authors used transformers to translate from chemical fingerprints to SMILES; showed that conversion re-construct the connectivity of the molecule
2
7
45
@gklambauer
Günter Klambauer
2 years
Please consider submitting to your work to our special issue "AI meets Toxicology"! We are happy to get submissions until November 30 (will remind you again ;)) !
Tweet media one
3
18
45
@gklambauer
Günter Klambauer
6 months
Researchers have been striving to bridge BIOLOGY 🧫 and CHEMISTRY ⚗️ via transcriptomics, metabolomics, and whatever.... ... funny that the bridge turned out to be the oldest biological technique: CELL MICROSCOPY.
0
8
44
@gklambauer
Günter Klambauer
1 year
Generative model for molecules in 3D space: using latent diffusions on point-structured latent spaces. As for all generative models,metrics is difficult: stability & molecule stability, validity, and validity & uniqueness used.
0
4
43
@gklambauer
Günter Klambauer
10 months
How good are large language models at answering questions about PATHWAYS, MOLECULAR INTERACTIONS, and MECHANISMS? Answer: Not too good, but the LLM trained on bio-literature (BioGPT) are better than the usual GPTs, etc.. Study:
0
9
42
@gklambauer
Günter Klambauer
11 months
Offline black-box optimization with diffusion models. The denoising process is conditioned on the label (output of black-box function). Also used for optimizing molecules (ChEMBL), but unclear how this was exactly done...
0
10
41
@gklambauer
Günter Klambauer
1 year
Based on seminal work of using Normalizing Flows to learn Boltzmann distributions, here is an approach that tries to alleviate the problem of the costly simulation of training data:
0
3
42
@gklambauer
Günter Klambauer
5 months
From @patwalters at ELLIS ML4Molecules: **STOP using MoleculeNet and TDC!** I fully support this statement (and have criticized those datasets myself for a long time)
4
5
41
@gklambauer
Günter Klambauer
1 year
Modeling long range intramolecular forces by extending message-passing networks: on addition to the local messages, nodes are mapped to non-local Fourier space and updated there (Ewald message passing). Improves MAE on energy in OC20:
0
7
41
@gklambauer
Günter Klambauer
7 months
Almost identical work to CLAMP "Contrastive Language-Assay-Molecule-Pretraining" (ICML2023): common space for natural language and mols. CLAMP allows to steer activity prediction with natural language models for zero-shot activity/property prediction.
Tweet media one
@Syd59067213
Haiteng Zhao
11 months
Introducing our new work: GIMLET: A Unified Graph-Text Model for Instruction-Based Molecule Zero-Shot Learning Say goodbye to the supervised molecule property prediction and embrace the instruction-based zero-shot paradigm with GIMLET!
Tweet media one
2
13
24
0
5
41
@gklambauer
Günter Klambauer
1 year
Since the LLM community has an evolutionary tree, I thought we should have one, too.. :) Sharing my personal perspective on the evolution of Deep Activitiy prediction Models (DAMs). PDF w links Bib
Tweet media one
@ylecun
Yann LeCun
1 year
A survey of LLMs with a practical guide and evolutionary tree. Number of LLMs from Meta = 7 Number of open source LLMs from Meta = 7 The architecture nomenclature for LLMs is somewhat confusing and unfortunate. What's called "encoder only" actually has an encoder and a decoder…
Tweet media one
74
742
3K
2
15
40
@gklambauer
Günter Klambauer
2 years
Self-supervised learning for molecular graphs is still in its infancy according to this paper: . For example, on the MUV dataset, the methods don't even beat a random method. Downstream: molecular property prediction. Suggest a new suite of probe tasks.
0
8
38
@gklambauer
Günter Klambauer
2 years
Transformer architecture learns chemical substructures together with natural language and can answer to questions about chemistry: (similar to our BioassayCLR method, but less focused on activity/property prediction).
1
1
39
@gklambauer
Günter Klambauer
7 months
We welcome the new researchers @sohvi_luukkonen and @weilincv to the Institute for Machine Learning and @LITAILab of @jkulinz !
Tweet media one
2
1
39
@gklambauer
Günter Klambauer
8 months
Wow, this is really cool: Contrastive pre-training on DNA sequences an the human genome to build a retrieval system. Then making a DNA vector database -- allows to align reads via maximum-inner-product-search!
1
3
40
@gklambauer
Günter Klambauer
8 months
I will be giving non-scientific, non-technical talks about #AI , #LargeLanguageModels , and #MachineLearning at the #ArsElectronicaFestival the next three days! Happy to discuss and meet you there!
Tweet media one
1
5
40