Page Not Found
Page not found. Your pixels are in another canvas.
A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.
Page not found. Your pixels are in another canvas.
This is a page not in th emain menu
Published:
This post will show up by default. To disable scheduling of future posts, edit config.yml
and set future: false
.
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
M Taquet, R Gong, SK Warfield. Organization for Human Brain Mapping (OHBM), 2013
RA Johnson, R Gong, S Greatorex-Voith, A Anand, A Fritzler. Bloomberg Data for Good Exchange Conference, 2015
R Gong, M Xie. Statistical Theory and Related Fields, 3(1), 35-36, 2019
R Gong. PMLR 103, 225-234, 2019
R Gong. arXiv:1909.12237, 2019
Y Chen, R Gong, M Xie. ArXiv:2001.08336, 2020
R Gong, JB Kadane, MJ Schervish, T Seidenfeld, RB Stern. ERGO (to appear), 2020
R Gong. ArXiv:2006.08522, 2020
R Gong and XL Meng. ACM-IMS Foundations of Data Science Conference (FODS-2020), 2020
R Gong, XL Meng. Statistical Science (with discussion, to appear), 2021
PE Jacob, R Gong, PT Edlefsen, AP Dempster. JASA (with discussion, to appear), 2021
Published:
Differential privacy (DP) is a mathematical framework that protects confidential information in a transparent and quantifiable way. I discuss how two classes of approximate computation techniques can be systematically adapted to produce exact statistical inference using DP data. For likelihood inference, we call for an importance sampling implementation of Monte Carlo expectation-maximization, and for Bayesian inference, an approximate Bayesian computation (ABC) algorithm suitable for possibly complex likelihood. Both approaches deliver exact statistical inference with respect to the joint statistical model inclusive of the differential privacy mechanism, yet do not require analytical access of such joint specification. Highlighted is a transformation of the statistical tradeoff between privacy and efficiency, into the computational tradeoff between approximation and exactness. Open research questions on two fronts are posed: 1) how to afford computationally accessible and (approximately) correct statistical analysis tools to DP data users; 2) how to understand and remedy the effect of any necessary post-processing with statistical analysis.
Published:
Algorithms are now widely used to streamline decisions in different contexts: insurance, health care, criminal justice. As some have shown, algorithms can make disproportionately more errors to the detriment of disadvantaged minorities compared to other groups. The literature in computer science has articulated different criteria of algorithmic fairness, each plausible in its own way. Yet, several impossibility theorems show that no algorithm can satisfy more than a few of these fairness criteria at the same time. We set out to investigate why this is so. In this talk, we first show that all criteria of algorithm fairness can be simultaneously satisfied under a peculiar and idealized set of premises. These include assumptions about access to information, representativeness of training data, capacity of the model, and crucially the construct of individual risk as the quantity to be assessed by the algorithm. When these assumptions are relaxed, we invoke a multi-resolution framework to understand the deterioration of the algorithm’s performance in terms of both accuracy and fairness. We illustrate our results using a suite of simulated studies. While our findings do not contradict existing impossibility theorems, they shed light on the reasons behind such failure and offer a path towards a quantitative and principled resolution. Joint work with Marcello Di Bello.
Published:
The Dempster-Shafer (DS) theory to statistical inference extends the Bayesian approach by allowing the use of partial and vacuous prior information. It yields three-valued uncertainty assessments representing probabilities for, against, and don’t know about formal assertions of interest. This talk presents a Gibbs algorithm that targets the distribution of a class of random convex polytopes, first described in Dempster (1972) to encapsulate the DS inference for Categorical distributions. Our sampler relies on an equivalence between the iterative constraints of the vertex configuration and the non-negativity of cycles in a fully connected directed graph. Joint work with Pierre E. Jacob, Paul T. Edlefsen and Arthur P. Dempster.
Graduate course, Rutgers University, 2020
Graduate course, Rutgers University, 2020
Byrne Seminar, Rutgers University, 2020
Graduate course, Rutgers University, 2021