Statistics Seminar - Spring 2026
Seminars are held on Thursdays from 1:00 - 2:00pm on Zoom unless otherwise noted. For access information, please contact the Math Department.
For questions about the seminar schedule, please contact Chenlu Shi and Yan Sun
*******************************************************************************************************
February 26
Dr. Chi-Kuang Yeh, Georgia State University
Single and multi-objective optimal designs for group testing experiments
Group testing, or pooled-sample testing, is widely used in large-scale screening and resource-constrained studies, yet principled design methodology for precise parameter estimation remains limited. This talk presents an optimal design framework for group testing that targets efficient estimation of key model parameters while accounting for cost constraints. Beyond classical criteria such as D-, D_s-, and A-optimality, a central novelty is the introduction of maximin design principles, including potentially non-differentiable criteria, into group testing procedures. These nondifferentiable criteria have not previously been explored in this context and yield designs with strong worst-case guarantees and improved robustness. The framework accommodates both large-sample settings through optimal approximate designs and small-sample studies through exact optimal designs, enabling systematic assessment of robustness to changes in criteria, statistical models, and cost structures. We demonstrate the practical impact of this approach through an application to Chlamydia screening using imperfect assays under budget constraints, and show that precise parameter estimation via optimal design is a foundational step that directly enables efficient and reliable sequential group-testing procedures.
This talk will be mainly based on the paper that is in revision, https://arxiv.org/abs/2508.08445, and the other two papers that are under review.
Homepage: https://chikuang.github.io/
*******************************************************************************************************
March 12
Dr. Aram-Alexandre Pooladian, Yale University
Blind denoising diffusion models and the blessings of dimensionality
We analyze, theoretically and empirically, the performance of generative diffusion models based on \emph{blind denoisers}, in which the denoiser is not given the noise amplitude in either the training or sampling processes. Assuming that the data distribution has low intrinsic dimensionality, we prove that blind denoising diffusion models (BDDMs), despite not having access to the noise amplitude, \emph{automatically} track a particular \emph{implicit} noise schedule along the reverse process. Our analysis shows that BDDMs can accurately sample from the data distribution in polynomially many steps as a function of the intrinsic dimension. Empirical results corroborate these mathematical findings on both synthetic and image data, demonstrating that the noise variance is accurately estimated from the noisy image. Remarkably, we observe that schedule-free BDDMs produce samples of higher quality compared to their non-blind counterparts. We provide evidence that this performance gain arises because BDDMs correct the mismatch between the true residual noise (of the image) and the noise assumed by the schedule used in non-blind diffusion models. This is joint work with Zahra Kadkhodaie, Sinho Chewi, and Eero Simoncelli.
Preprint available at https://arxiv.org/abs/2602.09639
Homepage: https://arampooladian.com/
*******************************************************************************************************
April 2
Tareq Aldirawi, NJIT
Conformal Risk Control Under Non-Monotone Loss
Conformal risk control (CRC) provides distribution-free guarantees for controlling the expected loss at a user-specified level. Existing theory typically assumes that the loss decreases monotonically with a tuning parameter that governs the size of the prediction set. This assumption is often violated in practice, where losses may behave non-monotonically due to competing objectives such as coverage and efficiency.
We study CRC under non-monotone loss functions when the tuning parameter is selected from a finite grid, a common scenario in thresholding or discretized decision rules. Revisiting a known counterexample, we show that the validity of CRC without monotonicity depends on the relationship between the calibration sample size and the grid resolution. In particular, risk control can still be achieved when the calibration sample is sufficiently large relative to the grid.
We provide a finite-sample guarantee for bounded losses over a grid of size $m$, showing that the excess risk above the target level $\alpha$ is of order $\sqrt{\log(m)/n}$, where $n$ is the calibration sample size. A matching lower bound shows that this rate is minimax optimal. We also derive refined guarantees under additional structural conditions, including Lipschitz continuity and monotonicity, and extend the analysis to settings with distribution shift via importance weighting.
Numerical experiments on synthetic multilabel classification and real object detection data illustrate the practical impact of non-monotonicity. Methods that account for finite-sample deviations achieve more stable risk control than approaches based on monotonicity transformations, while maintaining competitive prediction-set sizes.
*******************************************************************************************************
April 9
Dr. Xiang Li, University of Pennsylvania
What Can Statistics Offer to Language Models: Watermarking and Evaluation
Large language models (LLMs) have transformed how we generate and process information, yet two foundational challenges remain: ensuring the authenticity of their outputs and accurately evaluating their true capabilities. In this talk, I argue that both challenges are, at their core, statistical problems, and that statistical thinking can play an important role in advancing reliable and principled research on large language models. I will present two lines of work that approach these problems from a statistical perspective.
The first part introduces a statistical framework for language watermarks, which embed imperceptible signals into model-generated text for provenance verification. By formulating watermark detection as a hypothesis testing problem, this framework identifies pivotal statistics, provides rigorous Type I error control, and derives optimal detection rules that are both theoretically grounded and computationally efficient. It clarifies the theoretical limits of existing methods, such as the Gumbel-max watermark, and guides the design of more robust and powerful detectors. The second part focuses on language model evaluation, where I study how to quantify the unseen knowledge that models possess but may not reveal through limited queries. To that end, I introduce a statistical pipeline, based on the smoothed Good–Turing estimator, to estimate the total amount of a model’s knowledge beyond what is observed in benchmark datasets. The findings reveal that even advanced LLMs often articulate only a fraction of their internal knowledge, suggesting a new perspective on evaluation and model competence. Together, these projects represent an ongoing effort to develop statistical foundations for trustworthy and reliable language models, with applications ranging from watermark detection to model evaluation.
This talk is based on the following works:
https://arxiv.org/abs/2404.01245
https://arxiv.org/abs/2506.02058
and will briefly mention follow-up studies:
https://arxiv.org/abs/2411.13868
https://arxiv.org/abs/2510.22007
Homepage: https://lx10077.github.io/
*******************************************************************************************************
April 23
Dr. Yuchen Wu, Cornell University
Title/Abstract Forthcoming
Homepage: https://wuyc0114.github.io/
*******************************************************************************************************
April 30
Dr. Amir Sagiv, NJIT Department of Mathematical Sciences
Title/Abstract Forthcoming
Homepage: https://asagivmath.github.io/
*******************************************************************************************************
Last updated: April 10, 2026