Recent Publications


@inproceedings{Bascou_Lebre_Salmon20,
 author = {F. Bascou and S. Lèbre and J. Salmon},
 booktitle = {Journées de Statistique},
 pdf = {http://josephsalmon.eu/papers/JDS2020.pdf},
 title = {Debiasing the Elastic Net for models with interactions},
 year = {2020}
}

@inproceedings{Bertrand_Klopfenstein_Blondel_Vaiter_Gramfort_Salmon20,
 author = {Q. Bertrand and Q. Klopfenstein and M. Blondel and S. Vaiter and A. Gramfort and J. Salmon},
 booktitle = {ICML},
 comment = { [Code] },
 pdf = {https://arxiv.org/pdf/2002.08943.pdf},
 title = {Implicit differentiation of Lasso-type models for hyperparameter optimization},
 year = {2020}
}

@inproceedings{Bertrand_Klopfenstein_Blondel_Vaiter_Gramfort_Salmon20b,
 author = {Q. Bertrand and Q. Klopfenstein and M. Blondel and S. Vaiter and A. Gramfort and J. Salmon},
 booktitle = {CAP},
 comment = { [Code] },
 pdf = {https://arxiv.org/pdf/2002.08943.pdf},
 title = {Implicit differentiation of Lasso-type models for hyperparameter optimization},
 year = {2020}
}

@inproceedings{Massias_Bertrand_Gramfort_Salmon20,
 author = {Q. Bertrand and M. Massias and A. Gramfort and J. Salmon},
 booktitle = {AISTATS},
 pages = {2655--2665},
 pdf = {https://arxiv.org/pdf/2001.05401.pdf},
 title = {Support recovery and sup-norm convergence rates for sparse pivotal estimation},
 year = {2020}
}

@inproceedings{Chevalier_Gramfort_Salmon_Thirion20,
 author = {J.-A. Chevalier and A. Gramfort and J. Salmon and B. Thirion},
 booktitle = {NeurIPS},
 pdf = {https://arxiv.org/pdf/2009.14310.pdf},
 title = {Statistical control for spatio-temporal MEG/EEG source imaging with desparsified multi-task Lasso},
 year = {2020}
}

@techreport{Chevalier_Nguyen_Varoquaux_Salmon_Thirion20,
 author = {J.-A. Chevalier and T.-B. Nguyen and G. Varoquaux and J. Salmon and B. Thirion},
 journal = {ArXiv e-prints},
 title = {Decoding with Confidence: Statistical Control on Decoder Maps},
 year = {2020}
}

@article{Deledalle_Papadakis_Salmon_Vaiter19b,
 author = {C.-A. Deledalle and N. Papadakis and J. Salmon and S. Vaiter},
 journal = {J. Math. Imaging Vis.},
 pdf = {https://arxiv.org/pdf/1910.11186.pdf},
 title = {Block based refitting in $\ell_{12}$ sparse regularisation},
 year = {2020}
}

Full list of publications

Contact

Email: joseph.salmon@umontpellier.fr

Address: IMAG, c.c. 051
Université de Montpellier
Place Eugène Bataillon
34095 Montpellier Cedex 5
(office 415, building 9)

Phone: +33 4 67 14 35 19

News

Team

PhD Students

Alumni

Joining my team?

I am always looking for outstanding and highly motivated people to join my team as interns, PhD students, post-doctorates or research engineers in the following areas:

  • optimization for machine learning (including federated learning, privacy, etc.)
  • high dimensional and robust statistics

I always have open positions for outstanding applicants. The application process is light:

  1. Send me an email with your CV, transcript of most recent grades (for interns and PhD students) and explain in a paragraph why you are interested to join my group.
  2. After preliminary feedback on my side, I will ask you to secure two reference letters (one is enough for interns or PhD students) to be sent directly to me.
  3. At this stage an interview (possibly online) will be arranged to double check your skills and profile compatibility.

Short Bio

Since 2018, I am a full professor at Université de Montpellier and an associate member at INRIA Parietal Team. For the spring and summer quarters 2018, I was a visiting assistant professor at UW, Statistics departement. From 2012 to 2018 I was an assistant professor at Telecom ParisTech. Back in 2011 and 2012, I was a post-doctoral Associate at Duke university working with Rebecca Willett.

In 2010, I finished my PHD in statistics and image processing under the supervision of Dominique Picard and Erwan Le Pennec at the Laboratoire de Probabilités et de Modélisation Aléatoire, now LPSM, in Université Paris Diderot.

Software

  • BenchOpt: package to simplify, make more transparent and more reproducible comparisons between optimization algorithms

  • Celer: a fast Lasso solver, code of the associated ICML2018 paper "Dual Extrapolation for Faster Lasso Solvers" (pdf, slides)

  • sparse-ho: a fast hyper-parameter package to select the best Lasso parameter efficiently, code of the associated ICML2020 paper "Implicit differentiation of Lasso-type models for hyperparameter optimization" (pdf)
  • matlab toolboxes for statistics and image processing (this is legacy)

More on my Github Page

Teaching / courses

Full list of courses

Talks

Full list of talks, with slides and possibly videos when recorded.

Miscellaneous

Here (Miscellaneous), you will find some (math)art and other distractions.