Thesis: Expected Complexity and Gradients of Deep Maxout Neural Networks and Implications to Parameter Initialization. Supervisor: Prof. Guido Montúfar (Group Leader at MPI MiS and Professor at UCLA).
Degree awarded by Leipzig University. Research conducted at MPI MiS with parallel enrollment in IMPRS.
- Proposed a stable initialization method for deep maxout networks, achieving over 40% accuracy improvement; published in ICML.
- Proved that expected complexity grows polynomially with depth in maxout networks despite earlier assumptions of exponential growth; published in NeurIPS.
- Contributed to a study on loss landscapes of ReLU networks; published in TMLR.