A Comparison of Large Language Model and Human Performance on Random Number Generation Tasks

Rachel M. Harrison·August 19, 2024

Summary

A preliminary study compared the performance of ChatGPT-3.5, a large language model (LLM), and humans on Random Number Generation Tasks (RNGTs) to assess if ChatGPT-3.5 exhibits human-like cognitive biases when generating random number sequences. The study found that ChatGPT-3.5 more effectively avoided repetitive and sequential patterns compared to humans, with notably lower repeat frequencies and adjacent number frequencies. The research suggests that further investigation into different models, parameters, and prompting methodologies could help LLMs more closely mimic human random generation behaviors, expanding their applications in cognitive and behavioral science research. The study investigates LLMs' ability to mimic human cognitive processes, specifically in generating random sequences. Unlike traditional computational methods, LLMs process and generate sequences in a way that could potentially mirror human randomness, marked by inherent biases and imperfections. The research aims to compare LLMs' performance against human performance in random number generation tasks (RNGTs) to understand how generative AI might replicate and diverge from human cognitive biases. The study uses contemporary LLMs, focusing on OpenAI's ChatGPT model, to replicate existing human studies' conditions. The researchers crafted a specific user prompt to ensure the model's outputs reflect its standard operational parameters. The results show that ChatGPT demonstrates a higher level of apparent randomness than human outputs but still falls short of achieving perfect randomness. The study's findings contribute to a better understanding of LLMs' capabilities and limitations in generating random sequences. In conclusion, the study highlights the potential of large language models in generating random sequences that mimic human behavior, albeit with some limitations. The findings suggest that further research is needed to optimize LLMs' performance in this area, potentially expanding their applications in cognitive and behavioral science research. The study also underscores the importance of understanding the unique characteristics of LLM-generated sequences compared to human-generated ones, which could inform the design of new experimental methodologies for AI behavioral research.

Key findings

4

Introduction
Background
Overview of Random Number Generation Tasks (RNGTs)
Importance of RNGTs in cognitive and behavioral science research
Brief history of LLMs and their role in AI
Objective
To assess the performance of ChatGPT-3.5, a large language model, in generating random number sequences compared to humans
Investigate if ChatGPT-3.5 exhibits human-like cognitive biases in RNGTs
Method
Data Collection
Description of the experimental setup and conditions
Selection of participants (humans) and the ChatGPT-3.5 model
User prompts for ChatGPT-3.5 to generate random number sequences
Data Preprocessing
Data collection process for human-generated sequences
Data cleaning and preprocessing for ChatGPT-3.5 outputs
Comparison metrics for evaluating randomness and cognitive biases
Results
Performance Analysis
Comparison of repeat frequencies and adjacent number frequencies between human and ChatGPT-3.5 outputs
Statistical analysis of the results to highlight significant differences
Insights on Cognitive Biases
Examination of the extent to which ChatGPT-3.5 mimics human cognitive biases in RNGTs
Discussion on the limitations of ChatGPT-3.5 in achieving perfect randomness
Discussion
Limitations and Future Directions
Analysis of the study's findings in the context of LLMs' capabilities and limitations
Importance of understanding the unique characteristics of LLM-generated sequences
Recommendations for future research to optimize LLMs' performance in RNGTs
Applications in Cognitive and Behavioral Science
Potential applications of LLMs in cognitive and behavioral research
Expanding the scope of AI in understanding human cognitive processes
Conclusion
Summary of Findings
Recap of the study's main results and implications
Contribution to the understanding of LLMs' capabilities in generating random sequences
Implications for Research and Practice
Future research directions to enhance LLMs' performance in RNGTs
Practical applications of AI in cognitive and behavioral science research
Basic info
papers
computation and language
neurons and cognition
artificial intelligence
Advanced features