The NBER working paper “How People Use ChatGPT” (September, 2025) provides unprecedented insights into how ChatGPT is actually used globally.
Citation: Chatterji, A., Cunningham, T., Deming, D. J., Hitzig, Z., Ong, C., Shan, C. Y., & Wadman, K. (2025). How people use ChatGPT (NBER Working Paper No. 34255). National Bureau of Economic Research. https://doi.org/10.3386/w34255
Here are some takeaways and their methodology:
ChatGPT achieved extraordinary adoption : By July 2025, it reached 700 million weekly active users (10% of the global adult population) and processed 2.5 billion daily messages. This represents the fastest technology diffusion in history, surpassing even the internet's early adoption rates.
Personal use grew faster than work use : Non-work messages increased from 53% to 73% between June 2024 and June 2025. This challenges the dominant narrative that AI's primary economic impact comes through workplace productivity.
Consumer surplus is massive : The authors cite Collis and Brynjolfsson's estimate that US users would need $98 compensation to give up generative AI for a month, implying $97 billion in annual consumer surplus.
Nearly 80% of usage falls into three categories :
Programming is surprisingly small : Only 4.2% of ChatGPT messages involve coding, contradicting assumptions about AI's primary technical applications.
"Asking" dominates over "Doing" : 49% of messages seek advice/information ("Asking") versus 40% requesting task completion ("Doing"). Among work messages, Asking messages received higher quality ratings and grew faster.
Knowledge work applications : 81% of work-related messages involve "obtaining/interpreting information" or "making decisions and solving problems" - suggesting ChatGPT functions more as a research assistant than task automator.
Gender gap closed dramatically : From 80% male users in early 2023 to 52% female users by June 2025, representing complete reversal of the gender disparity.
Youth dominance : 46% of messages come from users under 26, though this concentration has slightly decreased over time.
Global expansion in lower-income countries : Adoption grew fastest in countries with $10,000-40,000 GDP per capita, indicating AI access is democratizing globally.
Education and occupation matter for work use : Users with graduate degrees are 48% likely to send work-related messages versus 37% for those without bachelor's degrees. Professional occupations show 50-57% work usage versus 40% for non-professional roles.
User satisfaction improved over time : "Good" interactions became 4x more common than "Bad" by July 2025, up from 3x in late 2024.
Writing tasks dominate professional use : 40% of work messages involve writing, with management/business users reaching 52%. This reflects writing as a universal white-collar skill.
Home production impact equals workplace impact : The dominance of non-work usage suggests AI's economic value extends far beyond measured workplace productivity into unmeasured household efficiency and welfare.
Decision-making enhancement in knowledge work : The prevalence of information-seeking and advisory use cases indicates ChatGPT primarily augments human judgment rather than replacing human tasks, particularly valuable in knowledge-intensive occupations where better decisions drive productivity.
These findings fundamentally reframe how we should think about AI's economic impact - less about job displacement and automation, more about enhanced decision-making and consumer welfare across both work and personal contexts.
The researchers used privacy-preserving methodology to analyze ChatGPT usage data. Here's how they did it:
1. Growth Dataset
2. Classified Messages Dataset
3. Employment Dataset
The critical innovation : No human researcher ever saw actual user messages. Instead, they used automated LLM classifiers:
Step 1: PII Removal
Step 2: Automated Classification
Step 3: Context Inclusion
Secure separation : Employment data held by external vendor, never directly accessed by researchers
Query restrictions :
Privacy controls : Individual records never visible, only statistical summaries above threshold
Human validation : Researchers validated classifiers using WildChat dataset (publicly available third-party chatbot conversations)
Agreement metrics :
Quality checks : Cross-referenced automated sentiment analysis with actual user thumbs-up/down feedback on 60,000 interactions
Representative sampling : Messages weighted to reflect actual daily volume patterns, not just random selection
Exclusion criteria consistently applied :
Strengths :
Limitations :
This methodology represents a significant advance in studying digital behavior while maintaining privacy. The automated classification approach could become a template for analyzing sensitive user data in other contexts.