Generative AI in Genomics (Gen2): Barriers and Frontiers

ICLR 2026 Workshop  |  April 27  |  Rio de Janeiro, Brazil (Hybrid)

Contact: genai-in-genomics@googlegroups.com


Tentative Workshop Program & Schedule


2026 Workshop Agenda

Date: Monday, April 27, 2026

Location: Room TBD, Riocentro Convention and Event Center


Time Duration Event
9:00 - 9:10 AM 10 mins Welcoming Remarks (Organizers)
9:10 - 10:10 AM 60 mins

Sarah Teichmann (University of Cambridge)

Keynote: Mapping the Human Body One Cell at a Time

Abstract: The 37 trillion cells of the human body have a remarkable array of specialised functions and must cooperate in time and space to construct a functioning human. Combining single cell and spatial genomics with AI/ML tools and data science, my lab has been attempting to understand this cellular diversity, how it is generated during development and how it goes wrong in disease. My talk will present recent advances towards a virtual Human Cell Atlas (HCA). I will highlight methods we developed to integrate and annotate cell atlases, and discuss how these models can be used to disentangle the effect of covariates, predict unseen perturbations and decode the regulatory logic of human cell types. I will additionally describe GenAI-based approaches we developed to model the spatial organization of cell types within organs, which can be leveraged to predict the effect of a cell on its neighbours.

10:10 - 10:50 AM 40 mins

Maria Brbic (EPFL)

TBD

Abstract: TBD

10:50 - 11:10 AM 20 mins Coffee Break
11:10 - 11:50 AM 40 mins

Nic Fishman (Harvard University)

Generative Distribution Embeddings

Abstract: Many real-world problems require reasoning across multiple scales, demanding models which operate not on single data points, but on entire distributions. We introduce generative distribution embeddings (GDE), a framework that lifts autoencoders to the space of distributions. In GDEs, an encoder acts on sets of samples, and the decoder is replaced by a generator which aims to match the input distribution. This framework enables learning representations of distributions by coupling conditional generative models with encoder networks which satisfy a criterion we call distributional invariance. We show that GDEs learn predictive sufficient statistics embedded in the Wasserstein space, such that latent GDE distances approximately recover the W2 distance, and latent interpolation approximately recovers optimal transport trajectories for Gaussian and Gaussian mixture distributions. We systematically benchmark GDEs against existing approaches on synthetic datasets, demonstrating consistently stronger performance. We then apply GDEs to six key problems in computational biology: learning representations of cell populations from lineage-tracing data (150K cells), predicting perturbation effects on single-cell transcriptomes (1M cells), predicting perturbation effects on cellular phenotypes (20M single-cell images), modeling tissue-specific DNA methylation patterns (253M sequences), designing synthetic yeast promoters (34M sequences), and spatiotemporal modeling of viral protein sequences (1M sequences).

11:50 - 12:40 PM 50 mins

Contributed Talks

Selected work from submissions

~3 talks selected from submitted papers. TBD.

12:40 - 1:10 PM 50 mins Lunch Break
1:10 - 2:10 PM 60 mins Poster Session #1 (Accepted submissions)
2:10 - 2:50 PM 40 mins

Gokcen Eraslan (Genentech)

TBD

Abstract: TBD

2:50 - 3:30 PM 40 mins

Adam Kosiorek (Google Deep Mind)

TBD

Abstract: TBD

3:30 - 4:30 PM 60 mins

Panel Discussion

Panelists:
TBD

4:30 - 4:45 PM 20 mins Coffee Break
4:45 PM - End Poster Session #2 (Accepted submissions)

Poster Sessions


Poster Assignment:

Details to follow upon acceptance.


Poster Presentation Format

  • More instructions and poster template (optional to use)to follow upon acceptance.

Poster Session Times:

  • Poster Session 1: 1:30 – 2:30 PM
  • Poster Session 2: 4:30 PM – End