Project in Prairie
As the amount of data available and the complexity of the models keep increasing, a number of issues arise in deploying optimization techniques for artificial intelligence at scale. Such challenges have long been integrated in high-performance computing, where the combination of optimization with other fields from numerical linear algebra to differential equations has led to powerful algorithms. This project aims at adopting a similar approach with optimization methods for data science.
My research aims at developing optimization methods for artificial intelligence that leverage existing methodology and advances from scientific computing along two axes. On one hand, we motivate the use of standard algorithmic frameworks for scientific computing in modern learning tasks by proposing practical schemes with complexity guarantees. Our research will aim at analyzing the complexity of classical second-order methods used in scientific computing so as to design frameworks with theoretical grounds and practical appeal for artificial intelligence. On the other hand, we develop derivative-free algorithms for automated parameter tuning of complex data science models. Our setting will be that of expensive, black-box systems for which a number of parameters require calibration.

