The chair develops novel optimization methods for machine learning, where one can exploit some structure in the learning pipeline. In particular, the omnipresence of discrete aspects –both of output objects, learned by models, and input objects, that we predict from– has arisen as a key property methods need to capture. Incorporating discrete structures when building and training models opens a broad set of challenges due to harder optimization problems. Among our end-goals, we want to build robustness against various forms of perturbations into such learning models by leveraging the additional structure.
ACTIVITIES
The SOL chair is starting in January 2026 and will be hiring two PhDs and one postdoc.
Published on November 13, 2025 Updated on November 18, 2025
Core members
Mathieu Besançon
Jérôme Malick
Quoc-Tung Le
Hamza Ennaji
Research topics
Optimization for learning, convex optimization, discrete optimization, machine learning, robustness
Share the linkCopyCopiedClose the modal windowShare the URL of this pageI recommend:Consultable at this address:La page sera alors accessible depuis votre menu "Mes favoris".Stop videoPlay videoMutePlay audioChat: A question? Chatbot Robo FabricaMatomo traffic statisticsX (formerly Twitter)