If you are interested in an undergraduate or graduate research project under my supervision, please reach out by email or in person. It is helpful if you include a brief description of your background and interests in your initial message.

Possible topics

My research projects will generally fall within the scope of  machine learning (and most likely deep learning). Particular interests of mine are for example the following:
  • Training of neural networks. How can we find good weights in our neural network, and if there are many good weights, which ones does a training algorithm find?
  • Depth separation. What kind of problems are better solved by deeper neural networks?
  • Convolutional neural networks. Why do CNNs outperform fully connected networks in image classification? 
  • Generalization of neural networks. Does a neural network uncover meaningful structure in a data set, or merely memorize the particular examples we showed it during training? 
If you have a particular topic in mind, I will be happy to see how it may be integrated. The range for undergraduate research projects is somewhat wider, including other topics in machine learning, but also the calculus of variations, partial differential equations, and geometric flows.


  • Real analysis. Most of my research involves analysis in some way, and familiarity with the results and techniques from real analysis in one and several dimensions is indispensable. This includes for example Sard's theorem, the regular value theorem, Lebesgue integration (including limit theorems), basic topology.
  • Probability theory. Knowledge of measure-based probability is preferred, familiarity with the basics (law of large numbers, central limit theorem) is required. Stochastic processes, conditional expectations and high-dimensional statistics are useful.

Useful Qualifications

The following are not required, nor is it realistic to be an expert in all of them. In any research project, it is likely that one or several of them may pop up at some point.
  • Programming experience. While I am happy to supervise theory-heavy projects, machine learning research often combines theory and practice. Knowledge of Python and specifically the TensorFlow or PyTorch libraries is particularly useful.
  • Functional analysis. Functional analysis is essentially the linear algebra of infinite-dimensional spaces. It pops up in many places and is a very useful background in many different places, often relating to function spaces.
  • Statistical learning theory. How do neural networks trained on finite data sets perform on unseen data? Particularly Rademacher complexities and concentration inequalities may be helpful.
  • Ordinary differential equations. Training algorithms of neural networks are modeled after ordinary differential equations, which are often useful to gain intuition. 
  • Partial differential equations. The training of infinitely wide neural networks can be described by partial differential equations. Conversely, it is an active problem how the tools of machine learning can be used in scientific computing, which often includes the solution of certain PDEs.
  • Differential geometry. Real data in applications seems to be concentrated on 'low-dimensional' subsets of a high-dimensional space. Understanding sub-manifolds of Euclidean spaces or even abstract Riemannian manifolds can be useful. 
  • Measure theory. Measure theory is another topic that pops up in many different places, including function representation and e.g. the geometry of low-dimensional sets which are too 'rough' to be captured by manifolds.
  • Stochastic analysis. Stochastic gradient descent and related algorithms can be modeled by stochastic differential equations in the vanishing step-size limit.