Eliciting Problem Specifications via Large Language Models

Robert E. Wray, James R. Kirk, John E. Laird·May 20, 2024

Summary

This paper explores the use of large language models (LLMs) in automating problem-solving by translating natural language descriptions into semi-formal specifications for cognitive systems. The authors design an LLM-enabled cognitive task analyst agent that generates problem-space definitions and strategies from AI literature. The goal is to reduce human intervention in problem formulation and enable AI systems to tackle diverse problems using domain-general methods. Key points include: 1. LLMs as a tool for automating problem representation, potentially speeding up AI research. 2. Problem spaces and cognitive task analysis (CTA) frameworks, like GOMS, are used to structure problem-solving processes. 3. The CTA agent generates problem specifications adhering to Polya's problem-solving principles and Newell's definition of problem spaces. 4. The agent aims to create problem formulations for weak methods, allowing for autonomous problem-solving. 5. Two alternative system designs are discussed, one relying heavily on LLMs and another with more human intervention. The study highlights the potential of LLMs in streamlining AI development and suggests future directions for multi-modal inputs and more sophisticated problem-solving capabilities. However, it also acknowledges the need for refining problem representations and addressing limitations in current models.

Key findings

5

Introduction
Background
Emergence of large language models in AI research
Challenges in problem formulation for AI systems
Objective
To explore LLMs for problem representation and CTA
Reduce human intervention in problem formulation
Enable AI autonomy with domain-general methods
Method
Data Collection
Literature review on problem-solving principles (Polya, Newell)
Analysis of cognitive task analysis frameworks (e.g., GOMS)
Data Preprocessing
Selection and preprocessing of relevant LLM data
Integration of problem-solving strategies into LLM models
LLM-Enabled Cognitive Task Analyst Agent
Model Architecture
Design of the agent for semi-formal specification generation
Comparison of LLM-based and human-in-the-loop systems
Problem Space Definition
Adherence to problem-solving principles
Application of Newell's problem space concept
Strategy Generation
Weak method identification and formulation
Automation of Polya's problem-solving steps
Performance Evaluation
Assessing accuracy and efficiency of generated specifications
Comparison with human-generated problem statements
Limitations and Future Directions
refining problem representations for better accuracy
Multi-modal inputs and LLM improvements
Addressing current model limitations for advanced problem-solving
Conclusion
LLMs' potential in streamlining AI development
Implications for AI autonomy and problem-solving scalability
Recommendations for future research in cognitive systems and LLMs.
Basic info
papers
computation and language
artificial intelligence
Advanced features
Insights
What frameworks, like GOMS, are used to structure problem-solving processes in the study?
How do LLMs contribute to automating cognitive task analysis in the paper's research?
What is the primary focus of the paper regarding large language models and problem-solving?
What are the key principles the CTA agent follows when generating problem specifications?