Home / Fellows / John Krumm

Fellows

John Krumm

University of Southern California, USA - (Writing Residency)
The Privacy Power of Vagueness
01 January 2024 - 31 January 2024
Others
FacebookTwitter

John Krumm graduated from the School of Computer Science at Carnegie Mellon University with a PhD in robotics and a thesis on texture analysis in images. He worked at the Robotics Center of Sandia National Laboratories in Albuquerque, New Mexico.
He is currently an associate director of the Integrated Media Systems Center in the Viterbi School of Engineering at the University of Southern California. His research focuses on understanding peoples' location and personal data privacy. He currently serves on the editorial board of IEEE Pervasive Computing Magazine. He is the chair of the executive committee of ACM SIGSPATIAL and part of the Science Advisory Committee of the Geospatial Science and Human Security Division at Oak Ridge National Laboratory.

In January 2024, he joins the IAS for a one-month writing residency.

Research Interests

Human mobility, personal data privacy, geospatial processing

The Privacy Power of Vagueness

A person may willingly reveal their age, gender, and home city to a third party company. However, this same person may be uncomfortable with the inferences that come from this data, such as their income, political preferences, income, and education level.

Regular people should understand what can be inferred about them from revelations of seemingly innocuous personal data. There is likely a simple, underlying theoretical foundation that makes this clear. For instance, we can look at data to understand how a person’s age can be used to infer their income. This is based on a simple joint probability distribution of age and income. In theory, there is a larger joint distribution, often approximated with a deep neural network, that gives the probabilistic relationship between tens of different, personal variables. Examining a distribution like this will help understand:

  • What can be inferred from a revelation of a few personal details?
  • How does this change if a person gives fuzzy answers, such as an age range instead of a specific age?
  • How are the inferences affected if a person lies about their personal data? What are the best lies in order to confuse the inference?
  • Some inferences are more sensitive than others, e.g. an inference about income may be more sensitive than an inference about sports teams preferences. How does the sensitivity of the inferences affect the sensitivity of the personal data revelations?
  • A company receiving personal data will likely not reveal the trade secrets of what it can infer. How can an individual still make smart choices about what to reveal given the individual's uncertainty about what else the company can infer?

There is a simple, underlying theory that can answer these questions, using joint probabilities, probabilistic expectation, and possibly information theory. This can be easy to illustrate with census data, which will demonstrate answers to the questions above. This project can lead to practical, real-life guidelines on the consequences and best practices for revealing personal data.

Key and Lastest Publications

Banovic, Nikola, and John Krumm. "Warming Up to Cold Start Personalization." Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 1, no. 4 (2018): 1-13. (link)

Krumm, John. "Sensitivity Analysis of Personal Location Disclosure." In 2022 23rd IEEE International Conference on Mobile Data Management (MDM), pp. 73-82. IEEE, 2022. (link)

Krumm, John. "Maximum entropy bridgelets for trajectory completion." In Proceedings of the 30th International Conference on Advances in Geographic Information Systems, pp. 1-8. 2022. (link)

 

New session of the "Paris IAS Ideas" talk series, with the participation of John Krumm, University of Southern California / Paris IAS Fellow
05 Jan 2024 15:00 -
05 Jan 2024 15:30,
Personal data privacy

30378
2023-2024