# Recent Publications

@inproceedings{Bascou_Lebre_Salmon21,
author = {F. Bascou and S. Lèbre and J. Salmon},
booktitle = {EGC},
pdf = {https://editions-rnti.fr/render_pdf.php?p=1002656},
title = {ElasticNet avec gestion des interactions et débiaisage},
year = {2021}
}

@article{Bertrand_Klopfenstein_Massias_Blondel_Vaiter_Gramfort_Salmon21,
author = {Q. Bertrand and Q. Klopfenstein and M. Massias and M. Blondel and S. Vaiter and A. Gramfort and J. Salmon},
eprint = {2105.01637},
journal = {arXiv e-prints},
pages = {arXiv:2105.01637},
pdf = {https://arxiv.org/pdf/2105.01637},
primaryclass = {stat.ML},
title = {{Implicit differentiation for fast hyperparameter selection in non-smooth convex learning}},
year = {2021}
}

@article{Chevalier_Nguyen_Varoquaux_Salmon_Thirion20,
author = {J.-A. Chevalier and T.-B. Nguyen and G. Varoquaux and J. Salmon and B. Thirion},
journal = {Neuroimage},
pages = {117921},
pdf = {https://www.sciencedirect.com/science/article/pii/S1053811921001981},
title = {Decoding with Confidence: Statistical Control on Decoder Maps},
year = {2021}
}

@article{Deledalle_Papadakis_Salmon_Vaiter21,
author = {C.-A. Deledalle and N. Papadakis and J. Salmon and S. Vaiter},
journal = {J. Math. Imaging Vis.},
pages = {216-236},
pdf = {https://arxiv.org/pdf/1910.11186.pdf},
title = {Block based refitting in $\ell_{12}$ sparse regularisation},
volume = {63},
year = {2021}
}

@inproceedings{Liu_Salmon_Harchaoui21,
author = {L. Liu and J. Salmon and Z. Harchaoui},
booktitle = {ICASSP},
pdf = {https://stat.uw.edu/sites/default/files/2019-07/tr652.pdf},
title = {Score-based change detection for gradient-based learning machines},
year = {2021}
}

@inproceedings{Bascou_Lebre_Salmon20,
author = {F. Bascou and S. Lèbre and J. Salmon},
booktitle = {Journées de Statistique},
pdf = {http://josephsalmon.eu/papers/JDS2020.pdf},
title = {Debiasing the Elastic Net for models with interactions},
year = {2020}
}

@inproceedings{Bertrand_Klopfenstein_Blondel_Vaiter_Gramfort_Salmon20,
author = {Q. Bertrand and Q. Klopfenstein and M. Blondel and S. Vaiter and A. Gramfort and J. Salmon},
booktitle = {ICML},
comment = { [Code] },
pdf = {https://arxiv.org/pdf/2002.08943.pdf},
title = {Implicit differentiation of Lasso-type models for hyperparameter optimization},
year = {2020}
}

Full list of publications

# Contact

Address: IMAG, c.c. 051
Université de Montpellier
Place Eugène Bataillon
34095 Montpellier Cedex 5
(office 415, building 9)

Phone: +33 4 67 14 35 19

# Team

### Joining my team?

I am always looking for outstanding and highly motivated people to join my team as interns, PhD students, post-doctorates or research engineers in the following areas:

• optimization for machine learning (including federated learning, privacy, etc.)
• high dimensional and robust statistics

I always have open positions for outstanding applicants. The application process is light:

1. Send me an email with your CV, transcript of most recent grades (for interns and PhD students) and explain in a paragraph why you are interested to join my group.
2. After preliminary feedback on my side, I will ask you to secure two reference letters (one is enough for interns or PhD students) to be sent directly to me.
3. At this stage an interview (possibly online) will be arranged to double check your skills and profile compatibility.

# Short Bio

Since 2018, I am a full professor at Université de Montpellier and an associate member at INRIA Parietal Team. For the spring and summer quarters 2018, I was a visiting assistant professor at UW, Statistics departement. From 2012 to 2018 I was an assistant professor at Telecom ParisTech. Back in 2011 and 2012, I was a post-doctoral Associate at Duke university working with Rebecca Willett.

In 2010, I finished my PHD in statistics and image processing under the supervision of Dominique Picard and Erwan Le Pennec at the Laboratoire de Probabilités et de Modélisation Aléatoire, now LPSM, in Université Paris Diderot.

# Software

• BenchOpt: package to simplify, make more transparent and more reproducible comparisons between optimization algorithms

• Celer: a fast Lasso solver, code of the associated ICML2018 paper "Dual Extrapolation for Faster Lasso Solvers" (pdf, slides)

• sparse-ho: a fast hyper-parameter package to select the best Lasso parameter efficiently, code of the associated ICML2020 paper "Implicit differentiation of Lasso-type models for hyperparameter optimization" (pdf)
• matlab toolboxes for statistics and image processing (this is legacy)

More on my Github Page

# Open positions in my lab

This list is fuzzy so please contact me directly for potential opportunities (post-doc, phd thesis, internship)

# Teaching / courses

Full list of courses

# Talks

Full list of talks, with slides and possibly videos when recorded.

# Miscellaneous

Here (Miscellaneous), you will find some (math)art and other distractions.