Financial Times highlights Olivier David Zerbib’s research among the most “positively” cited studies


In a recent Financial Times article, the newspaper looks at how research is often measured by citation counts and journal rankings. To show how these metrics shape perceptions of impact, the article highlights the ten most-cited studies according to the Scite analytics tool — including one by Olivier-David Zerbib, researcher at CREST.

Read the Financial Times article here.

CRESTive Minds : Enjeux économiques et politiques des sanctions économiques


Clément Montes est doctorant au CREST-École polytechnique depuis 2022 sous la supervision d’Alessandro Riboni. Sa recherche se situe dans le domaine de la géo-économie et des échanges internationaux, avec un focus sur les sanctions économiques.

Sa thèse porte sur les sanctions économiques dites smart qui ciblent spécifiquement des entreprises. À partir de la typologie des sanctions proposée par la Global Sanction DataBase (GSDB), recensant les mesures adoptées à l’échelle mondiale, on constate que ce type d’instrument a été largement mobilisé au cours de la dernière décennie. Cependant, les bases de données existantes ne fournissent pas d’informations détaillées sur les cibles à l’intérieur d’un pays. Cette limite rend difficile l’évaluation de l’impact des sanctions, faute de connaître avec précision le secteur d’activité et les caractéristiques des entreprises visées.

Clément a ainsi constitué une nouvelle base de données recensant les entreprises spécifiquement sanctionnées par l’Union européenne (UE). Dans un premier temps, cette base, nommée Firm Sanction DataBase (FSDB), enrichit la liste accessible sur le site de l’UE en ajoutant des variables telles que la localisation des entités et le calendrier des sanctions (e.g. menace, début, pause, fin de la sanction). En agrégeant les informations de la FSDB, on retrouve les principales tendances observées dans la GSDB. Ces nouvelles informations mettent en évidence trois biais récurrents dans l’estimation de l’impact d’une sanction économique.

  1. Supposer que toutes les entités visées par une sanction contre un pays se trouvent effectivement sur son territoire conduit à une mauvaise classification d’environ 23% des entités répertoriées dans la FSDB. Cette hypothèse, fréquemment utilisée dans la littérature en raison du manque de données, peut donc biaiser les analyses.
  2. Ignorer les programmes de sanction dont les noms ne correspondent pas à un pays revient à ignorer les sanctions de 21% de la FSDB. Par exemple, les programmes relatifs à la cybersécurité, aux droits humains ou au terrorisme ne permettent pas d’associer directement une sanction à un pays donné, et ne sont donc pas étudiés. Néanmoins, ces programmes concernent également des entreprises situées dans les cibles habituelles de l’Union européenne.
  3. Postuler que les entreprises prennent connaissance de leur sanction le jour de son entrée en vigueur revient à négliger leurs anticipations stratégiques. En réalité, l’écart temporel entre la menace de sanction et sa mise en application laisse en moyenne deux mois aux firmes pour ajuster leur comportement. Cet intervalle peut aller jusqu’à un an. Ce constat est encore plus marqué dans le cas des sanctions secondaires : si l’on suppose qu’une entreprise anticipe sa propre sanction dès lors qu’un de ses partenaires commerciaux est visé, la période d’anticipation est en moyenne de un an et demi, mais peut s’étendre jusqu’à douze ans.

Dans un second temps, la FSDB enrichit la liste publique de 1,900 entreprises actuellement sanctionnées en ajoutant environ 1,500 entités supplémentaires, identifiées dans les archives de l’UE, dont les sanctions ont depuis été levées. La FSDB recense également les justifications associées à l’instauration et, le cas échéant, à la levée des sanctions pour chaque entité. Ces variables permettent ainsi d’analyser les motivations politiques sous-jacentes à chaque régime de sanctions.

L’analyse du vocabulaire des textes justifiant la levée des sanctions des 1,500 entités délistées montre que cette levée traduit, dans une certaine mesure , le succès de la coercition exercée sur la cible. Ainsi, le retrait d’une entreprise d’un programme de sanctions reflète une amélioration des résultats politiques du programme considéré, à l’échelle macroéconomique.

Figure 1 - Évolution journalière du taux de succès des sanctions européennes par objectifs politiques

Note : La figure illustre la proportion d’entreprises retirées parmi l’ensemble des entités sanctionnées pour un objectif donné. Une même entité peut être visée par plusieurs objectifs politiques. Ainsi, une entité sanctionnée pour deux objectifs est, par exemple, comptabilisée pour moitié dans chacun d’eux.

La FSDB permet d’analyser la proportion d’entreprises retirées de chaque programme et que l’on peut interpréter comme un indicateur du degré de réussite des sanctions. La figure 1 illustre l’évolution dynamique de cet indice. Le graphique met en évidence que certains objectifs politiques, tels que la défense des droits humaines ou la lutte contre le terrorisme, semblent produire des résultats politiques plus significatifs que ceux visant à endiguer des conflits armés ou à protéger la cybersécurité européenne.

Référence

Felbermayr, G., Kirilakha, A., Syropoulos, C., Yalcin, E., & Yotov, Y. V. (2020). The global sanctions data base. European Economic Review129, 102561.

Cet article s’inscrit dans l’évènement Les Rencontres Cybersécurité et Défense de l’Institut Polytechnique de Paris organisé le 21 octobre 2025. Un évènement proposé en partenariat avec le Centre Interdisciplinaire d’Études pour la Défense et la Sécurité (CIEDS) soutenu par le ministère des Armées et l’Agence de l’innovation de défense, ainsi que la chaire Cyber et souveraineté numérique – IHEDN, rattachée à l’Institut des Hautes Études de Défense Nationale à travers son Fonds de dotation.

Dette publique et stabilité financière : Giovanni Ricco éclaire les enjeux du TPI dans L’Express


Dans un article publié par L’Express le 15 octobre 2025, Giovanni Ricco, économiste au CREST et à l’École polytechnique, revient sur le rôle du TPI (Transmission Protection Instrument), un mécanisme mis en place par la Banque centrale européenne (BCE) pour éviter les tensions excessives sur les marchés de la dette.

Giovanni Ricco explique que ce dispositif permet à la BCE d’intervenir pour corriger certaines dynamiques de marché injustifiées, susceptibles de fragiliser la soutenabilité budgétaire des États membres. Il rappelle que cet instrument, créé en 2022, s’ajoute à la boîte à outils de la BCE pour assurer la bonne transmission de la politique monétaire dans l’ensemble de la zone euro.

« En période de tensions sur les marchés, les investisseurs peuvent rechercher la sécurité de certaines obligations, ce qui fait grimper le rendement d’autres pays », souligne-t-il, insistant sur l’importance d’éviter des spirales auto-entretenues entre hausse des taux et dégradation budgétaire.

📖 À lire dans L’Express :
👉 Dette : le TPI, l’arme méconnue de la BCE qui protège la France (Thibault Marotte, 15 octobre 2025)

Anticiper les crues de la Seine : les travaux de Samuel Rufat mis à l’honneur


Un article publié dans Libération revient sur les travaux de Samuel Rufat, géographe au CREST et spécialiste des risques urbains, consacrés à la simulation d’une crue majeure de la Seine à Paris.

🧭 Ses recherches visent à mieux comprendre les effets potentiels d’un tel événement et à identifier les marges de progression en matière de préparation et de résilience urbaine. À travers des modélisations et des enquêtes de terrain, Samuel Rufat étudie la façon dont les habitants perçoivent le risque et s’y préparent, ainsi que le rôle des infrastructures de protection existantes.

💡 Ces travaux rappellent l’importance d’une culture du risque partagée, alliant prévention, aménagement du territoire et sensibilisation des citoyens, pour renforcer la capacité collective à faire face aux aléas climatiques.

Étienne Ollion invited to discuss France’s political moment in European media


As France experiences a new wave of political turbulence, Étienne Ollion, sociologist at CREST-CNRS and École Polytechnique, has been invited by several major European media outlets to help make sense of the current situation.

Over the past week, he has shared his insights on Radio France’s Questions du Soir, Austria’s Ö1 – Europa JournalDenik N in the Czechia, and Politiken, one of Denmark’s leading newspapers. All of them turned to him to better understand how French institutions and public opinion are reacting to the latest government resdhuffle and budget debates.

Throughout these interviews, Étienne Ollion offers a sociologist’s perspective on the moment: how political uncertainty connects to deeper transformations in the evolution of party systems, and how current responses are out of sync with citizens’ expectations of public action.

He notes that France’s tripolar political landscape, divided between the presidential camp, the left, and the far right, has made coalition-building and compromise increasingly difficult in Parliament.

On Ö1 – Europa Journal, Austria’s leading public radio channel for culture and information, he explained that the rapid turnover of governments reflects not only political tensions but also “a deeper transformation of democratic life, where citizens expect more responsiveness and clarity from public action, and a wider role in the making of the nation”.

In Denik N, he pointed to growing frustration across social groups and a steady decline in confidence toward political representatives. And on France Culture, during a debate asking “Can France still be governed?”, he discussed the ways out of the political stalemate.

In Politiken, he provided international readers with context on how France’s political structures and social expectations interact in this volatile moment.

These interventions are closely linked to Étienne Ollion’s research on political representation and the evolution of elites. In his book The Candidates (Oxford Press, 2024), he studied how newcomers to politics navigate institutions and hierarchies, shedding light on the renewal and professionalization of political life. His more recent work, Une étrange victoire. L’extrême droite contre la politique (Seuil, 2024, with Michaël Foessel), explores the underpinnings of the rise of the far-right, between the growing opposition to democratic representation, a changing media landscape, and the new communication strategy of its main party, the National Rally.

At CREST, he develops research using computational methods to analyse public debate and political representation. This mix of empirical study, institutional knowledge, and data-driven analysis makes him a valuable voice for understanding the forces currently reshaping French democracy.

19 CREST Papers Accepted at NeurIPS 2025


This year again, CREST research team have made a strong mark at one of the world’s most prestigious conferences in artificial intelligence and machine learning: NeurIPS 2025.

A total of 19 papers by CREST members have been accepted, reflecting the lab’s growing influence in the global research community and the vitality of its teams in statistics, finance-insurance, and data science.

The accepted papers showcase the breadth and depth of the work carried out within the CREST.

Several contributions advance the fast-evolving field of diffusion and generative models, while others deepen the theoretical understanding of optimal transport and Wasserstein-based methods, approaches that connect CREST’s expertise in statistics and quantitative finance.

Other works explore kernel and inference methods, bandit algorithms, or privacy-preserving learning, illustrating the CREST’s commitment to developing both fundamental and responsible AI.

Beyond their diversity, these projects share a common spirit: a blend of mathematical rigor, computational innovation, and interdisciplinary collaboration. Together, they embody CREST’s ongoing mission to push the boundaries of what data-driven research can achieve, from theory to real-world impact.

Authors: Louis Allain, Sébastien Da Veiga, Brian Staber

Abstract: Conformal Prediction (CP) is a popular framework for constructing prediction bands with valid coverage in finite samples, while being free of any distributional assumption. A well-known limitation of conformal prediction is the lack of adaptivity, although several works introduced practically efficient alternate procedures. In this work, we build upon recent ideas that rely on recasting the CP problem as a statistical learning problem, directly targeting coverage and adaptivity. This statistical learning problem is based on reproducible kernel Hilbert spaces (RKHS) and kernel sum-of-squares (SoS) methods. First, we extend previous results with a general representer theorem and exhibit the dual formulation of the learning problem. Crucially, such dual formulation can be solved efficiently by accelerated gradient methods with several hundreds or thousands of samples, unlike previous strategies based on off-the-shelf semidefinite programming algorithms. Second, we introduce a new hyperparameter tuning strategy tailored specifically to target adaptivity through bounds on test-conditional coverage. This strategy, based on the Hilbert-Schmidt Independence Criterion (HSIC), is introduced here to tune kernel lengthscales in our framework, but has broader applicability since it could be used in any CP algorithm where the score function is learned. Finally, extensive experiments are conducted to show how our method compares to related work. All figures can be reproduced with the accompanying code.

Link: https://doi.org/10.48550/arXiv.2505.21039

Authors: Marguerite Petit–Talamon, Marc Lambert, anna Korba

Abstract:Variational inference (VI) is a popular approach in Bayesian inference, that looks for the best approximation of the posterior distribution within a parametric family, minimizing a loss that is typically the (reverse) Kullback-Leibler (KL) divergence. In this paper, we focus on the following parametric family: mixtures of isotropic Gaussians (i.e., with diagonal covariance matrices proportional to the identity) and uniform weights. We develop a variational framework and provide efficient algorithms suited for this family. In contrast with mixtures of Gaussian with generic covariance matrices, this choice presents a balance between accurate approximations of multimodal Bayesian posteriors, while being memory and computationally efficient. Our algorithms implement gradient descent on the location of the mixture components (the modes of the Gaussians), and either (an entropic) Mirror or Bures descent on their variance parameters. We illustrate the performance of our algorithms on numerical experiments.

Link: https://doi.org/10.48550/arXiv.2506.13613

Authors: Touqeer Ahmad, Mohammadreza Mousavi Kalan, François Portier, Gilles Stupfler

Abstract: Oversampling synthetic minority examples using and its variants is a leading strategy for addressing imbalanced classification problems. Despite the success of this approach in practice, its theoretical foundations remain underexplored. We develop a theoretical framework to analyze the behavior of and related methods when classifiers are trained on synthetic data. First, we establish an exponential inequality that characterizes the gap between the empirical risk computed on synthetic samples and the true population risk on the minority class. Second, we show that a kernel-based classification rule trained on synthetic data can achieve the minimax rate of convergence. This leads to practical guidelines for better parameter tuning of both and the downstream learning algorithm. Numerical experiments are provided to illustrate and support the theoretical findings.

Link: https://neurips.cc/virtual/2025/poster/117246

Authors: Vahan Arsenyan, Elen Vardanyan, Arnak Dalalyan

Abstract: Generative modeling aims to produce new random examples from an unknown target distribution, given access to a finite collection of examples. Among the leading approaches, denoising diffusion probabilistic models (DDPMs) construct such examples by mapping a Brownian motion via a diffusion process driven by an estimated score function. In this work, we first provide empirical evidence that DDPMs are robust to constant-variance noise in the score evaluations. We then establish finite-sample guarantees in Wasserstein-2 distance that exhibit two key features: (i) they characterize and quantify the robustness of DDPMs to noisy score estimates, and (ii) they achieve faster convergence rates than previously known results. Furthermore, we observe that the obtained rates match those known in the Gaussian case, implying their optimality.

Link: https://doi.org/10.48550/arXiv.2506.09681

Authors: Marius Potfer, Vianney Perchet

Abstract: Repeated multi-unit auctions, where a seller allocates multiple identical items over many rounds, are common mechanisms in electricity markets and treasury auctions. We compare the two predominant formats: uniform-price and discriminatory auctions, focusing on the perspective of a single bidder learning to bid against stochastic adversaries. We characterize the learning difficulty in each format, showing that the regret scales similarly for both auction formats under both full-information and bandit feedback, as and , respectively. However, analysis beyond worst-case regret reveals structural differences: uniform-price auctions may admit faster learning rates, with regret scaling as   in settings where discriminatory auctions remain at . Finally, we provide a specific analysis for auctions in which the other participants are symmetric and have unit-demand, and show that in these instances a similar regret rate separation appears.

Link: https://neurips.cc/virtual/2025/poster/115489

Authors: Georgios Gavrilopoulos, Guillaume Lecué, Zong Shang

Abstract: We obtain upper bounds for the estimation error of Kernel Ridge Regression (KRR) for all non-negative regularization parameters, offering a geometric perspective on various phenomena in KRR. As applications: 1. We address the multiple descent problem, unifying the proofs of arXiv:1908.10292 and arXiv:1904.12191 for polynomial kernels and we establish multiple descent for the upper bound of estimation error of KRR under sub-Gaussian design and non-asymptotic regimes. 2. For a sub-Gaussian design vector and for non-asymptotic scenario, we prove a one-sided isomorphic version of the Gaussian Equivalent Conjecture. 3. We offer a novel perspective on the linearization of kernel matrices of non-linear kernel, extending it to the power regime for polynomial kernels. 4. Our theory is applicable to data-dependent kernels, providing a convenient and accurate tool for the feature learning regime in deep learning theory. 5. Our theory extends the results in arXiv:2009.14286 under weak moment assumption.
Our proof is based on three mathematical tools developed in this paper that can be of independent interest: 1. Dvoretzky-Milman theorem for ellipsoids under (very) weak moment assumptions. 2. Restricted Isomorphic Property in Reproducing Kernel Hilbert Spaces with embedding index conditions. 3. A concentration inequality for finite-degree polynomial kernel functions.

Link: https://doi.org/10.48550/arXiv.2404.07709

Authors: Yvann Le Fay, Nicolas Chopin, Simon Barthelmé

Abstract: Variational inference consists in finding the best approximation of a target distribution within a certain family, where `best’ means (typically) smallest Kullback-Leiber divergence. We show that, when the approximation family is exponential, the best approximation is the solution of a fixed-point equation. We introduce LSVI (Least-Squares Variational Inference), a Monte Carlo variant of the corresponding fixed-point recursion, where each iteration boils down to ordinary least squares regression and does not require computing gradients. We show that LSVI is equivalent to stochastic mirror descent; we use this insight to derive convergence guarantees. We introduce various ideas to improve LSVI further when the approximation family is Gaussian, leading to a O(d³) complexity in the dimension d of the target in the full-covariance case, and a O(d)complexity in the mean-field case. We show that LSVI outperforms state-of-the-art methods in a range of examples, while remaining gradient-free, that is, it does not require computing gradients.

Link: https://doi.org/10.48550/arXiv.2502.18475

Authors: Luca Arnaboldi, Bruno Loureiro, Ludovic Stephan, Florent Krzakala, Lenka Zdeborova

Abstract: We study the dynamics of stochastic gradient descent (SGD) for a class of sequence models termed Sequence Single-Index (SSI) models, where the target depends on a single direction in input space applied to a sequence of tokens. This setting generalizes classical single-index models to the sequential domain, encompassing simplified one-layer attention architectures. We derive a closed-form expression for the population loss in terms of a pair of sufficient statistics capturing semantic and positional alignment, and characterize the induced high-dimensional SGD dynamics for these coordinates. Our analysis reveals two distinct training phases: escape from uninformative initialization and alignment with the target subspace, and demonstrates how the sequence length and positional encoding influence convergence speed and learning trajectories. These results provide a rigorous and interpretable foundation for understanding how sequential structure in data can be beneficial for learning with attention-based models.

Link: https://doi.org/10.48550/arXiv.2506.02651

Authors: Oussama Zekri, Nicolas Boullé

Abstract: Discrete diffusion models have recently gained significant attention due to their ability to process complex discrete structures for language modeling. However, fine-tuning these models with policy gradient methods, as is commonly done in Reinforcement Learning from Human Feedback (RLHF), remains a challenging task. We propose an efficient, broadly applicable, and theoretically justified policy gradient algorithm, called Score Entropy Policy Optimization (SEPO), for fine-tuning discrete diffusion models over non-differentiable rewards. Our numerical experiments across several discrete generative tasks demonstrate the scalability and efficiency of our method. Our code is available at this https URL.

Link: https://doi.org/10.48550/arXiv.2502.01384

Authors: Stanislas Strasman, Sobihan Surendran, Claire Boyer, Sylvain Le Corff, Vincent Lemaire, Antonio Ocello

Abstract: Score-based Generative Models (SGMs) have achieved impressive performance in data generation across a wide range of applications and benefit from strong theoretical guarantees. Recently, methods inspired by statistical mechanics, in particular Hamiltonian dynamics, have introduced Critically-damped Langevin Diffusions (CLD), which define diffusion processes in extended spaces by coupling the data with auxiliary variables. These approaches, along with their associated score-matching and sampling procedures, have been shown to outperform standard diffusion-based samplers numerically. In this paper, we propose an upper bound on the sampling error for CLD-based generative models in the Wasserstein metric. To better exploit the extended space, we also propose a modified dynamic that introduces an additional hyperparameter controlling the noise applied to the data coordinates. This hyperparameter influences the smoothness of sample paths, and our discretization error analysis offers practical guidance for tuning, leading to improved sampling performance.

Link: https://neurips.cc/virtual/2025/poster/117301

Authors: Nina Vesseron, Louis Béthune, Marco Cuturi

Abstract: The canonical approach in generative modeling is to split model fitting into two blocks: define first how to sample noise (e.g. Gaussian) and choose next what to do with it (e.g. using a single map or flows). We explore in this work an alternative route that ties sampling and mapping. We find inspiration in moment measures, a result that states that for any measure ρ, there exists a unique convex potential u such that ρ=ueu. While this does seem to tie effectively sampling (from log-concave distribution eu) and action (pushing particles through u), we observe on simple examples (e.g., Gaussians or 1D distributions) that this choice is ill-suited for practical tasks. We study an alternative factorization, where ρ is factorized as wew, where w is the convex conjugate of a convex potential w. We call this approach conjugate moment measures, and show far more intuitive results on these examples. Because w is the Monge map between the log-concave distribution ew and ρ, we rely on optimal transport solvers to propose an algorithm to recover w from samples of ρ, and parameterize w as an input-convex neural network. We also address the common sampling scenario in which the density of ρ is known only up to a normalizing constant, and propose an algorithm to learn w in this setting.

Link: https://doi.org/10.48550/arXiv.2503.10576

Author: Imad Aouali

Abstract: Efficient exploration is a key challenge in contextual bandits due to the large size of their action space, where uninformed exploration can result in computational and statistical inefficiencies. Fortunately, the rewards of actions are often correlated and this can be leveraged to explore them efficiently. In this work, we capture such correlations using pre-trained diffusion models; upon which we design diffusion Thompson sampling (dTS). Both theoretical and algorithmic foundations are developed for dTS, and empirical evaluation also shows its favorable performance.

Link: https://doi.org/10.48550/arXiv.2402.10028

Authors: Jung-Hun Kim, Milan Vojnovic, Min-hwan Oh

Abstract: We study the combinatorial semi-bandit problem where an agent selects a subset of base arms and receives individual feedback. While this generalizes the classical multi-armed bandit and has broad applicability, its scalability is limited by the high cost of combinatorial optimization, requiring oracle queries at *every* round. To tackle this, we propose oracle-efficient frameworks that significantly reduce oracle calls while maintaining tight regret guarantees. For worst-case linear rewards, our algorithms achieve regret using only oracle queries. We also propose covariance-adaptive algorithms that leverage noise structure for improved regret, and extend our approach to general (non-linear) rewards. Overall, our methods reduce oracle usage from linear to (doubly) logarithmic in time, with strong theoretical guarantees.

Link: https://neurips.cc/virtual/2025/poster/119751

Authors: Rémi Castera, Felipe Garrido, Patrick Loiseau, Simon Mauras, Mathieu Molina, Vianney Perchet

Abstract: We consider matroid allocation problems under opportunity fairness constraints: resources need to be allocated to a set of agents under matroid constraints (which includes classical problems such as bipartite matching). Agents are divided into C groups according to a sensitive attribute, and an allocation is opportunity-fair if each group receives the same share proportional to the maximum feasible allocation it could achieve in isolation. We study the Price of Fairness (PoF), i.e., the ratio between maximum size allocations and maximum size opportunity-fair allocations. We first provide a characterization of the PoF leveraging the underlying polymatroid structure of the allocation problem. Based on this characterization, we prove bounds on the PoF in various settings from fully adversarial (wort-case) to fully random. Notably, one of our main results considers an arbitrary matroid structure with agents randomly divided into groups. In this setting, we prove a PoF bound as a function of the size of the largest group. Our result implies that, as long as there is no dominant group (i.e., the largest group is not too large), opportunity fairness constraints do not induce any loss of social welfare (defined as the allocation size). Overall, our results give insights into which aspects of the problem’s structure affect the trade-off between opportunity fairness and social welfare.

Link: https://arxiv.org/pdf/2403.00397v2

Authors: Shiyun Lin, Simon Mauras, Nadav Merlis, Vianney Perchet

Abstract: We study the problem of matching markets with ties, where one side of the market does not necessarily have strict preferences over members at its other side. For example, workers do not always have strict preferences over jobs, students can give the same ranking for different schools and more. In particular, assume w.l.o.g. that workers’ preferences are determined by their utility from being matched to each job, which might admit ties. Notably, in contrast to classical two-sided markets with strict preferences, there is no longer a single stable matching that simultaneously maximizes the utility for all workers.
We aim to guarantee each worker the largest possible share from the utility in her best possible stable matching. We call the ratio between the worker’s best possible stable utility and its assigned utility the \emph{Optimal Stable Share} (OSS)-ratio. We first prove that distributions over stable matchings cannot guarantee an OSS-ratio that is sublinear in the number of workers. Instead, randomizing over possibly non-stable matchings, we show how to achieve a tight logarithmic OSS-ratio. Then, we analyze the case where the real utility is not necessarily known and can only be approximated. In particular, we provide an algorithm that guarantees a similar fraction of the utility compared to the best possible utility. Finally, we move to a bandit setting, where we select a matching at each round and only observe the utilities for matches we perform. We show how to utilize our results for approximate utilities to gracefully interpolate between problems without ties and problems with statistical ties (small suboptimality gaps).

Link: https://doi.org/10.48550/arXiv.2411.03270

Authors: Ziyad Benomar, Romain Cosson, Alexander Lindermayr, Jens Schöter

Abstract: In non-clairvoyant scheduling, the goal is to minimize the total job completion time without prior knowledge of individual job processing times. This classical online optimization problem has recently gained attention through the framework of learning-augmented algorithms. We introduce a natural setting in which the scheduler receives continuous feedback in the form of progress bars: estimates of the fraction of each job completed over time. We design new algorithms for both adversarial and stochastic progress bars and prove strong competitive bounds. Our results in the adversarial case surprisingly induce improved guarantees for learning-augmented scheduling with job size predictions. We also introduce a general method for combining scheduling algorithms, yielding further insights in scheduling with predictions. Finally, we propose a stochastic model of progress bars as a more optimistic alternative to conventional worst-case models, and present an asymptotically optimal scheduling algorithm in this setting.

Link: https://doi.org/10.48550/arXiv.2509.19662

Authors: Achraf Azize, Yulian Wu, Junya Honda, Francesco Orabona, Shinji Ito, Debabrota Basu

Abstract: As sequential learning algorithms are increasingly applied to real life, ensuring data privacy while maintaining their utilities emerges as a timely question. In this context, regret minimisation in stochastic bandits under ϵ-global Differential Privacy (DP) has been widely studied. Unlike bandits without DP, there is a significant gap between the best-known regret lower and upper bound in this setting, though they “match” in order. Thus, we revisit the regret lower and upper bounds of ϵ-global DP algorithms for Bernoulli bandits and improve both. First, we prove a tighter regret lower bound involving a novel information-theoretic quantity characterising the hardness of ϵ-global DP in stochastic bandits. Our lower bound strictly improves on the existing ones across all ϵ values. Then, we choose two asymptotically optimal bandit algorithms, i.e. DP-KLUCB and DP-IMED, and propose their DP versions using a unified blueprint, i.e., (a) running in arm-dependent phases, and (b) adding Laplace noise to achieve privacy. For Bernoulli bandits, we analyse the regrets of these algorithms and show that their regrets asymptotically match our lower bound up to a constant arbitrary close to 1. This refutes the conjecture that forgetting past rewards is necessary to design optimal bandit algorithms under global DP. At the core of our algorithms lies a new concentration inequality for sums of Bernoulli variables under Laplace mechanism, which is a new DP version of the Chernoff bound. This result is universally useful as the DP literature commonly treats the concentrations of Laplace noise and random variables separately, while we couple them to yield a tighter bound.

Link: https://doi.org/10.48550/arXiv.2505.05613

Authors: Marc Jourdan, Achraf Azize

Abstract: Best Arm Identification (BAI) algorithms are deployed in data-sensitive applications, such as adaptive clinical trials or user studies. Driven by the privacy concerns of these applications, we study the problem of fixed-confidence BAI under global Differential Privacy (DP) for Bernoulli distributions. While numerous asymptotically optimal BAI algorithms exist in the non-private setting, a significant gap remains between the best lower and upper bounds in the global DP setting. This work reduces this gap to a small multiplicative constant, for any privacy budget . First, we provide a tighter lower bound on the expected sample complexity of any δ-correct and -global DP strategy. Our lower bound replaces the Kullback–Leibler (KL) divergence in the transportation cost used by the non-private characteristic time with a new information-theoretic quantity that optimally trades off between the KL divergence and the Total Variation distance scaled by . Second, we introduce a stopping rule based on these transportation costs and a private estimator of the means computed using an arm-dependent geometric batching. En route to proving the correctness of our stopping rule, we derive concentration results of independent interest for the Laplace distribution and for the sum of Bernoulli and Laplace distributions. Third, we propose a Top Two sampling rule based on these transportation costs. For any budget , we show an asymptotic upper bound on its expected sample complexity that matches our lower bound to a multiplicative constant smaller than 8. Our algorithm outperforms existing δ-correct and -global DP BAI algorithms for different values of .

Link: https://openreview.net/forum?id=IFso8G8gwJ

Authors: Adrien Vacher, Omar Chehab, Anna Korba

Abstract: Sampling from multi-modal distributions is challenging, even in low dimensions. We provide the first sampling algorithm for a broad class of distributions — including all Gaussian mixtures — with a query complexity that is polynomial in the parameters governing multi-modality, assuming fixed dimension. Our sampling algorithm simulates a time-reversed diffusion process, using a specific Monte Carlo estimator of the intermediate score functions. Unlike previous works, it avoids metastability, requires no prior knowledge of the mode locations, and does not rely on restrictive smoothness assumptions that exclude general Gaussian mixtures. We illustrate this result on a low-dimensional but challenging multi-modal sampling task, where our algorithm significantly outperforms existing approaches.

Link: https://neurips.cc/virtual/2025/poster/119085

Comment mettre la finance au service du climat ?


La finance verte est mise à l’épreuve, secoués par des flots politiques et macroéconomiques défavorables. Mais la dynamique n’est pas morte, et il est encore possible de la soutenir, estiment Peter Tankov et Olivier David Zerbib.

Tribune Les Echos du 18/09/2025

A New Academic Year, New Faces at CREST


Every year, CREST welcomes a new cohort of researchers who bring fresh perspectives, curiosity, and energy to our community. This fall, we are pleased to host new researchers, postdoctoral fellow, PhD students, and visiting researchers from around the world.

Their arrival marks an important moment in the life of our lab: the opportunity to grow, to renew dialogues across disciplines, and to continue building a vibrant and diverse research environment. Whether they join us for a few months or several years, these newcomers contribute to the intellectual richness and collaborative spirit that define CREST.

We are pleased to introduce the researchers who are joining CREST this year, each bringing their unique background and expertise to our four research areas: economics, sociology, statistics, and finance-insurance.

Researchers

This year, several new researchers are joining CREST, strengthening our core faculty and bringing new perspectives across ou disciplines.

Philippe Coulangeon is a CNRS Research Director, affiliated with the Centre for Research on Social Inequalities (CRIS) at SciencesPo Paris.

His work is set at the intersection of cultural sociology and social inequalities. He explores the stratification of cultural practices and tastes, the dynamics of mass versus elite culture, artistic professions, and the democratization of culture. More recently, he has integrated environmental concerns into his analysis.

More information here.

Fanny Landaud is a CNRS researcher from CY Cergy Paris University and she is an IZA Research Fellow since 2019.

Her research covers applied microeconomics – particularly labor, education, family, and health economics – with a focus on the determinants and consequences of socioeconomic and gender inequalities in education and the labor market.

More information here.

Antonio Ocello was a Postdoctoral Researcher in Statistics and Machine Learning at École polytechnique (CMAP) since 2023.

His research focuses on mean-field games and mean-field control problems, branching diffusion processes, and stochastic optimal control, combining advanced probabilistic methods with applications in machine learning. He has contributed to work on convergence bounds for trust-region policy optimization in mean-field games and is developing stochastic target frameworks for branching processes.

More information here.

Enrico Ruolino was a Senior Researcher at the University of Lausanne.

His work focuses on public and labor economics, with a particular emphasis on behavioral responses to tax policy, gender inequality, intergenerational mobility, and educational outcomes.

More information here.

Clémentine Van Effenterre was previously an Assistant Professor of Economics at the University of Toronto.

Her research lies at the intersection of labor economics, applied microeconomics, and political economy. Notably, she investigates how norms, institutions, and policies shape labor market outcomes – examining issues like maternal labor supply, gender gaps in science and tech, and how having daughters influences political attitudes. She also hosts the “InequaliTalks” podcast on economics and inequality.

More information here.

Postdoctoral Fellows

Our new postdoctoral researchers are contributing to ongoing projects and launching their own work in close collaboration with CREST teams.

Stéphane Lhaut, in the Finance-Insurance team, with Caroline Hillairet and Olivier Lopez.

Angelica Martinez-Leyva, in the Economics cluster at CREST-ENSAI, with Marion Goussé

Andrea Pandolfi, in Statistics, with Nicolas Chopin

Arthur Stephanovitch, in Statistics, with Austin Stromme

Lishu Zhang, in the Finance-Insurance team, with Olivier-David Zerbib

PhD Students

This year, we welcome a new cohort of PhD students who are beginning their research journeys within our vibrant academic community.

30 new PhD students are joining our laboratory in our different research clusters.

9 PhD students will join our Economic team, 7 in the Finance-Insurance cluster, 12 new PhD students in Statistics, and 2 in our Sociology team.

4 of them will join our lab on the CREST-ENSAI campus in Bruz while the remaining PhD students will join our CREST-ENSAE Paris campus in Palaiseau.

This new team will also be joined by 3 PhD students in ATER contracts in the CREST-ENSAI campus.

Visiting Researchers

CREST is also hosting several visiting researchers this year, joining us for research stays that foster international collaboration.

Yang Chen (Lausanne University, Switzerland) is joining our Economic team to work with Julien Combe on Microeconomic theory and matching theory.

Matias Ortiz (Universidad de Chile, Chile) will work with Vianney Perchet on Learning-Augmented Online Algorithms for Matroid-Constrained problems under Partial in our Statistical team.

Simon Luck (Universita di Bologna, Italy) is joining Etienne Ollion’s team in Sociology to work on using natural language processing to study the relevance of the news media for political representation and decision-making processes.

Jules Verin (ENS Lyon, France) will work with Bertrand Garbinti in Economics on Wealth Accumulation, Life Cycle, and Taxation.

Väinö Yrjänäinen (Uppsala University, Sweden) is joining the Sociology team with Etienne Ollion to work on using work embeddings and transformer models in computational social sciences.