A Kernel Two-Sample Test, Arthur Gretton, Karsten M. Borgwardt, Malte J. Rasch, Bernhard Schölkopf, and Alex Smola, 2012Journal of Machine Learning Research, Vol. 13DOI: 10.5555/2810848.2810903 - Introduces the Maximum Mean Discrepancy (MMD) as a non-parametric kernel-based method for two-sample testing, which is central to quantitative multivariate comparison.
Generative Adversarial Networks, Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio, 2014Advances in Neural Information Processing Systems, Vol. 27 (NeurIPS)DOI: 10.48550/arXiv.1406.2661 - Presents the original concept of Generative Adversarial Networks (GANs), foundational for discriminator-based evaluation of synthetic data quality.
A Survey of Metrics for Generative Models, Zhifeng Wang, Jun Zou, 2020ACM Computing Surveys, Vol. 53 (ACM)DOI: 10.1145/3403986 - A comprehensive review of metrics for evaluating generative models, covering various quantitative methods for assessing synthetic data quality, including fidelity and diversity aspects.
Elements of Information Theory, Thomas M. Cover and Joy A. Thomas, 2006 (Wiley) - A standard textbook on information theory, essential for understanding metrics like Kullback-Leibler (KL) and Jensen-Shannon (JS) Divergence used in distribution comparisons.