Vector representations of generative models and their consistent estimation

Michael W. Trosset Co-Author
Indiana University Bloomington
 
Carey E. Priebe Co-Author
Johns Hopkins University
 
Hayden S. Helm Co-Author
Helivan Research
 
Aranyak Acharyya First Author
Johns Hopkins University
 
Aranyak Acharyya Presenting Author
Johns Hopkins University
 
Tuesday, Aug 5: 11:50 AM - 12:05 PM
1219 
Contributed Papers 
Music City Center 
Generative models, like large language models or text-to-image diffusion models, can generate a random output or response after being given a query from a user. Representing them with vectors in a finite-dimensional Euclidean space based on their responses to a set of queries, facilitates statistical decision-making tasks on black-box generative models using conventional tools. We establish sufficient conditions for consistent estimation of population-level vector representations of a set of generative models based on their sample responses to a set of queries.

Keywords

generative models

multidimensional scaling

raw stress embedding 

Main Sponsor

Section on Statistical Learning and Data Science