Anastasiia Makarova

Anastasiia Makarova

PhD student

ETH Zurich

I am a final-year PhD student in Computer Science at ETH Zurich under the supervision of Andreas Krause and member of Learning and Adaptive Systems Group. I am also associated with the ETH AI Center. I work on sequential decision-making and representation learning for structured data, such as ​point clouds or graphs.

Research interests
The main question that motivates my research is: How can we actively learn new complex environments? My interests span multiple topics around sequential decision-making, bayesian optimization, representation learning, generative modeling. I am sparked by designing robust algorithms with quantified uncertainty and well-understood limitations (theoretically and empirically), that would be applicable in society-critical areas.

Prior to ETH, I studied Maths, Physics, and Data Science at MIPT and Skoltech – Moscow universities with great people and great courses. I worked on deep learning for weakly-supervised semantic segmentation under the supervision of Victor Lempitsky and interned at Columbia University under the co-supervision of Hod Lipson. I did ML internships at Yandex (deep learning for precipitation nowcasting) and Amazon Web Services (Bayesian optimization for AutoML).

  • MSc in Data Analysis, 2017

    Skolkovo Institute of Science and Technology (Skoltech), Moscow

  • MSc in Applied Mathematics (with honors), 2017

    Moscow Institute of Physics and Technology (Phystech), Moscow

  • BSc in Applied Math and Physics (with honors), 2015

    Moscow Institute of Physics and Technology (Phystech), Moscow



(2023). Adversarial Causal Bayesian Optimization.

Cite PDF

(2023). Safe Risk-averse Bayesian Optimization for Controller Tuning.

Cite PDF

(2022). Model-based Causal Bayesian Optimization. Spotlight at International Conference on Learning Representations (ICLR) 2023.

Cite Oral talk PDF Code Slides

(2022). Automatic Termination for Hyperparameter Optimization. Best Paper Award and Contributed Talk at International Conference on Automated Machine Learning (AutoML-Conf) 2022, ICLR Workshop on Neural Architecture Search 2021.

Cite Best paper award talk PDF Code Video Slides Blog post

(2021). Risk-averse Heteroscedastic Bayesian Optimization. Conference on Neural Information Processing Systems (NeurIPS).

Cite PDF Video Slides

(2021). Cherry-Picking Gradients: Learning Low-Rank Embeddings of Visual Data via Differentiable Cross-Approximation. International Conference on Computer Vision (ICCV).

Cite PDF Code Poster

(2020). Mixed-Variable Bayesian Optimization. International Joint Conference on Artificial Intelligence (IJCAI).

Cite PDF Short Video Long Video

(2020). Hierarchical Image Classification Using Entailment Cone Embeddings. CVPR Workshop on on Differential Geometry (DiffCVML), resulted into the publicly available mobile app BioDex.

Cite PDF Slides Code App


Research Intern
Google DeepMind
Aug 2023 – Apr 2023 Zurich
Research Intern working on RL for LLMs.
ML Intern
Jun 2020 – Oct 2020 Berlin
Research Intern at AWS AI working on Automatic Model Tuning and SageMaker Autopilot.
PhD Student / Research Assistant
Oct 2017 – Present Zurich

Research areas: Probabilistic Machine Learning, Bayesian Optimization, Deep Learning, Tensors, Computer Vision

Advisor: Prof. Andreas Krause

Research Intern
Jun 2016 – Aug 2016 Moscow
Research Intern at Yandex.Weather working on precipitation nowcasting using deep convolutional and recurrent models
Visiting Researcher
Dec 2016 – Sep 2016 New York

Worked with Hod Lipson and collaboration with Victor Lempitsky.

  • Worked on deep learning based methods for weakly-supervised semantic segmentation
  • Developed an efficient architecture for image-based plant disease detection


Supervised theses

I (co-)supervised MSc theses of several bright students, some resulting into research publications:

  • Alicja Chaszczewicz: Following Gradients to Calibrate Equilibrium Reaching Simulators,
    jointly with Max Paulus, ETH Zurich, December 2020 - May 2021.
  • Ankit Dhall: Learning Representations for Images With Hierarchical Labels (paper @DiffCVML'20),
    jointly with Octavian Ganea and Dario Pavllo, ETH Zurich, March - September 2019.
  • Erik Daxberger: Mixed-Variable Bayesian Optimization (paper IJCAI'20),
    jointly with Matteo Turchetta, ETH Zurich, October 2018 - April 2019.
  • Stefan Beyeler: Multi-fidelity Batch Bayesian Optimization for the Calibration of Transport System Simulations,
    jointly with Matteo Turchetta, ETH Zurich, October 2017 - April 2018.