Show Menu
Cheatography

Problem Solving, Decision Making, Reasoning Cheat Sheet by

problem solving definition, processes, heuristics, etc

1. Definition of a Problem in Cognitive Psychology

A problem is defined as any situation in which a person has a goal but does not immedi­ately know the best way to reach it.
It involves a gap between the current state and the desired goal state, with no obvious path to bridge the gap.

Key Components of a Problem
Initial State: The current, unsati­sfa­ctory situation.
Goal State: The desired outcome or solution.
Obstacles: The limita­tions or constr­aints that prevent easy movement from the initial to the goal state.
Operators: The available actions or tools that can be used to move toward the goal.

Criteria That Make Something a “Problem”:
Lack of an immediate solution: If the answer is obvious or automatic, it is not a problem in the cognitive sense.
Requires cognitive effort: The individual must think, strate­gize, or analyze.
Goal-d­ire­cte­dness: There must be an intended outcome or objective.
Involves decisi­on-­making and uncert­ainty: Solutions are not always clear-cut.

Algorithms in Cognitive Psychology

Definition of Algorithm
An algorithm is a system­atic, rule-based procedure used to solve a problem. It involves a sequence of operations that, when followed correctly, guarantees a correct solution.
In cognitive psycho­logy, algorithms are studied as one of the structured methods humans may (rarely) use to solve well-d­efined problems.
Key Properties of Algorithms
Step-b­y-step: Each action follows logically from the previous one.
Exhaus­tive: All possible pathways are considered.
Rule-g­ove­rned: Operates under fixed, pre-de­ter­mined rules.
Soluti­on-­gua­ran­teed: If a solution exists, the algorithm will find it.
Often comput­ati­onally expensive: May require a lot of time and mental effort.

Role of Algorithms in Human Cognition
While algorithms are infallible in theory, humans do not always use them due to cognitive limita­tions like attention, working memory load, or time pressure.
Noneth­eless, algorithms are important in modeling cognitive processes such as logical reasoning, mathem­atical problem solving, and scientific thinking.

Related Concepts:

Cognitive load: Mental effort needed to solve a problem.
Insight: Sudden realiz­ation of a solution.
Problem repres­ent­ation: The way a problem is mentally structured can affect ease of solving.

Algorithms vs Heuristics in Cognitive Psychology

Heuristics are mental shortcuts. Unlike algori­thms, they do not guarantee a correct solution but are faster and often used by humans in real-world decisions.
Cognitive psycho­logists often compare algorithms (ideal proble­m-s­olving) with heuristics (actual strategies people use).

Theore­tical Frameworks and Research

Newell and Simon’s Inform­ati­on-­Pro­cessing Approach (1972):
Introduced the General Problem Solver, a computer program simulating algori­thmic reasoning. It illust­rated how humans could hypoth­eti­cally approach problems algori­thm­ically, though in practice they often used heuristics.
Herbert Simon’s Bounded Ration­ality:
Highli­ghted that while algorithms represent ideal ration­ality, humans operate within cognitive limits. This makes purely algori­thmic thinking rare in day-to-day decision-making.
Cognitive Load Theory (Sweller):
Emphasizes that high working memory demand reduces the likelihood of using algori­thmic methods unless the individual is highly practiced.

Gestalt Psychology (Early Founda­tions):
While not focused on algori­thms, the Gestal­tists emphasized insight in problem solving—an altern­ative to stepwise logic. This contrast laid early groundwork for comparing algorithms with non-linear proble­m-s­olving.

Summary (Algor­ithms)

Algorithms represent the ideal of rational, structured problem solving. They are essential to unders­tanding how problem solving could work in optimal cognitive systems. However, due to human limita­tions, algorithms are often replaced by quicker, intuitive heuristics in real-world situat­ions. Noneth­eless, they remain central to modeling cognitive processes and developing AI systems.

Insight Learning

Definition
Insight learning refers to the sudden realiz­ation of a problem’s solution without the use of trial-­and­-error. It involves a cognitive reorga­niz­ation of inform­ation leading to an “Aha!” or “Eureka” moment.
Theorist: Wolfgang Köhler
Köhler was a Gestalt psycho­logist who emphasized that perception and unders­tanding are holistic. His work with chimpa­nzees laid the foundation for insight as a distinct form of learning.
Key Charac­ter­istics
Sudden­ness: The solution appears abruptly rather than through gradual attempts.
Perceptual Reorga­niz­ation: The problem is viewed in a new way, revealing the solution.
No Overt Trial-­and­-Error: Unlike Thornd­ike’s animals, subjects do not randomly try different methods.
Transfer of Learning: Once insight is achieved, it can be applied to similar problems.

Köhler’s Experi­ments
Conducted on chimpa­nzees in the Canary Islands.
In one study, a banana was placed out of reach, and chimpa­nzees used sticks or stacked boxes to retrieve it.
The animals did not solve the problem by repeated random attempts; instead, they paused and then acted with purpose, suggesting cognitive restructuring.

Cognitive Explan­ation
Involves accessing previously unconn­ected elements in memory and restru­cturing them.
The solution often comes after a period of incubation — a temporary break from conscious problem-solving.
Insight is associated with higher­-order cognitive functions such as abstra­ction, pattern recogn­ition, and divergent thinking.

Relevance to Cognitive Psychology
Supports the idea that learning is not always linear or behavi­orally observable.
Provides evidence against purely behavi­orist models of learning.
Related to creative thinking, complex problem solving, and real-life innova­tion.
Neuroi­maging studies show right hemisphere involv­ement (espec­ially anterior temporal lobe) during insight.
Insight is now studied alongside intuitive decisi­on-­making and creativity research.
Applic­ations
Educat­ional strategies that promote deep unders­tanding over memorization.
Problem solving in design thinking, innova­tion, therapy, and scientific discovery.
Used to explain sudden clarity in proble­m-based learning enviro­nments.
Contrast with Other Learning Models
Vs. Trial-­and­-Error Learning: Insight does not involve repeated failure before success.
Vs. Operant Condit­ioning: Insight is not reinforced increm­ent­ally; it emerges through internal proces­sing.

Types of Heuristics

Core Types
Availa­bility Heuristic
Judging the likelihood of an event based on how easily examples come to mind.
Example: Overes­tim­ating plane crashes after seeing news coverage.
Repres­ent­ati­veness Heuristic
Evaluating probab­ilities by comparing how similar an instance is to a prototype.
Example: Assuming someone is a librarian because they are quiet and introverted.
Anchoring and Adjustment Heuristic
Making estimates by starting from an initial value (anchor) and adjusting, often insufficiently.
Example: Being influenced by the first price offered in a negotiation.
Other Common Heuristics
Recogn­ition Heuristic
Preferring options that are recognized over those that are not, especially when knowledge is limited.
Simulation Heuristic
Judging the likelihood of an event based on how easily one can imagine it happening.
Affect Heuristic
Making decisions based on emotional responses rather than detailed analysis.
Fluency Heuristic
Assuming that inform­ation processed more fluently (e.g., read more easily) is more accurate or important.

Means-Ends Analysis

Defini­tion:
Means-Ends Analysis (MEA) is a proble­m-s­olving strategy used to reduce the difference between a current situation and a desired goal by breaking the problem into smaller subgoals.
Origin:
Developed by Newell and Simon in the 1950s.
Based on the idea that people solve problems by identi­fying differ­ences between the present state and the goal state.
Core Idea:
Compare the current state with the goal state.
Identify the biggest difference.
Choose an action (means) to reduce that difference.
If the action can’t be applied directly, set a new subgoal to achieve conditions that allow the action.
Repeat the process until the goal is reached.
Steps in Means-Ends analysis
Identify the current state.
Identify the goal state.
Determine the differ­ence(s) between the two.
Select the most signif­icant difference.
Find an operator (action) to reduce that difference.
If the operator can’t be applied, create a subgoal to make it applicable.
Apply the operator and update the current state.
Repeat the steps until the goal is achieved.
Example
Problem: You want to bake a cake, but you have no eggs.
Current state: No eggs.
Goal state: Have a cake.
Differ­ence: Missing eggs.
Operator: Go to the store and buy eggs.
Subgoal: Get money, go to store.
Apply operator, return with eggs.
Now you can bake the cake.
Advantages
Helps structure problem-solving.
Breaks down complex problems into manageable parts.

Limita­tions
Can be ineffi­cient if the subgoals are not well chosen.
Assumes the problem solver can correctly identify and apply operators.
This method of problem solving comes under the inform­ation processing approach to problem solving

General Problem Solver (GPS)

Defini­tion:
The General Problem Solver (GPS) is a computer program developed in the 1950s to simulate human problem-solving.
It uses logical steps and rules to solve well-d­efined problems by mimicking human cognitive strate­gies.
Developed By:
Allen Newell
Herbert A. Simon
J.C. Shaw
(1957)
Purpose:
To model how humans solve problems.
To serve as a universal proble­m-s­olving engine for AI and psychology research.
How GPS Works:
Define the problem (initial state, goal state, and rules).
Analyze the difference between the current and goal states.
Select an operator to reduce the difference.
If the operator can't be used, set a subgoal to make it usable.
Apply the operator and update the current state.
Repeat until the goal is reached.
Strengths:
First program to separate proble­m-s­olving method from problem content.
Helped lay the foundation for symbolic AI.
Modeled human-like reasoning.
Limita­tions
Could only solve well-s­tru­ctured problems (with clear rules and goals).
Not effective for real-world or ill-st­ruc­tured problems.
Required a lot of predefined inform­ation.

Analogical Problem Solving

Defini­tion:
Analogical problem solving is a strategy where a person solves a new problem (target problem) by referring to a previously solved problem (source problem) that is struct­urally similar.

It involves mapping relati­onships from the known to the unknown.
Key Steps in Analogical Problem Solving
Noticing a Relational Similarity
Recogn­izing that the current problem is similar to one you’ve seen before.

Retrieving a Source Problem
Recalling a past situation that resembles the current one.

Mapping Corres­ponding Elements
Aligning the structure of the old problem with the new one.
Identi­fying which elements play similar roles.

Applying the Mapping
Using the solution from the old problem to address the new problem.

Classic Experiment Example
Gick & Holyoak (1980s) – The Radiation Problem
Partic­ipants were given a difficult medical problem.
If previously told a struct­urally similar story (attacking a fortress with small forces from different sides), they were more likely to solve it.
Key finding: Analogical transfer improves when people are explicitly told to compare stories.
Types of Analogies
Surface analogy: Similar in details but not in structure.
Structural analogy: Similar in underlying relati­onship — more useful for problem solving.
Why It’s Important
Promotes creative problem solving.
Helps transfer knowledge across domains.
Essential in learning, reasoning, and intell­igence.
Strengths
Encourages flexible thinking
Aids in solving novel or unfamiliar problems
Builds on past experience and knowledge
Limita­tions
People often focus on surface features, not deeper structure
May fail if analogy is inappr­opriate or misleading
Requires prior experience with relevant problems
 

Types of Problems in Cognitive Psycho­logy:

Well-D­efined Problems:
Clear initial state, goal, and rules (e.g., solving a math equation).
Ill-De­fined Problems:
Ambiguous or unclear goals and solutions (e.g., designing a career plan).

Stages of Problem Solving

Problem Identi­fic­ation
Recogn­izing that a problem exists.
Distin­gui­shing between the current situation and the desired goal.
Requires attention, percep­tion, and sometimes intuition.
Example: Realizing you can't submit an assignment because your file is corrupted.
Problem Repres­ent­ation (or Unders­tanding the Problem)
Mentally organizing the elements of the problem.
Involves creating a “problem space” with possible states and transitions.
Good repres­ent­ation often simplifies the problem.
Example: Drawing a diagram or making a flowchart to visualize relati­ons­hips.
Strategy Formul­ation
Deciding how to approach the problem.
Choosing between strategies like trial and error, heuris­tics, or algorithms.
Involves planning, goal-s­etting, and sometimes setting subgoals.
Organi­zation of Inform­ation
Sorting relevant and irrelevant data.
Grouping inform­ation based on patterns, catego­ries, or importance.
Helps reduce cognitive load and improve focus.
Resource Allocation
Assessing time, energy, attention, and tools required.
Determ­ining how much effort or what external help might be needed.
Example: Deciding whether to solve the problem now or postpone it for later.
Monitoring (or Progress Tracking)
Contin­uously checking if the strategy is working.
Adjusting methods or correcting errors along the way.
Metaco­gnition (thinking about one’s own thinking) plays a big role here.
Evaluation (or Reviewing the Outcome)
Reflecting on the solution: did it work?
Assessing the outcome against the original goal.
Learning from mistakes and successes to improve future problem solving.

Cognitive Processing Involved:

Repres­ent­ation: Mental model or schema of the problem.
Planning and strate­gizing: Selecting and organizing steps.
Monito­ring: Keeping track of progress.
Evalua­tion: Judging if the goal is met or if another approach is needed.
Examples of Problems in Cognitive Contexts:
Solving a jigsaw puzzle (well-defined)
Choosing a college major (ill-defined)
Figuring out how to fix a broken device without instructions

Why It Matters in Cognitive Psycho­logy:

Problem solving is a core cognitive function that reveals how we learn, reason, and adapt.
Unders­tanding what consti­tutes a problem helps in designing cognitive tests and therap­eutic interv­ent­ions.

Key Theorists:

Allen Newell & Herbert Simon – Inform­ati­on-­pro­cessing approach to problem solving.
Karl Duncker – Insight and functional fixedness.
Gestalt Psycho­logists – Emphasis on perception and restru­cturing in problem solving.

Use of Algorithms in Problem Solving

Algorithms are most useful for:
Well-d­efined problems: These have a clear goal, starting point, and rules (e.g., solving a quadratic equation).
Tasks with limited variables: Such as number­-based puzzles or rule-based logic problems.

They are less effective for:
Time-s­ens­itive situations: Where fast approx­ima­tions are needed over perfect solutions.

Ill-de­fined problems: Where goals or paths are vague (e.g., resolving interp­ersonal conflict).

Types of Algorithms

Brute-­force search: Tries every possible option until the right one is found. Effective but inefficient.
Means-end analysis: Compares current state with goal state and takes steps to reduce the differ­ence. Common in both algori­thmic and heuristic frameworks.
Recursive algorithms: Solves a problem by breaking it down into smaller instances of the same problem. Concep­tually aligned with problem decomposition.
Search algorithms in memory: Used in modeling retrieval (e.g., serial exhaustive search).

Relevance to Cognitive Science and AI

In cognitive science and artificial intell­igence, algorithms are crucial for simulating proble­m-s­olving processes. Cognitive archit­ectures like ACT-R and SOAR are built around rule-based processing models that mimic algori­thmic thinking.
These models provide insight into how humans could solve problems if they followed strict comput­ational logic.

Cognitive Conditions for Algorithm Use

Formal education and training: Increases famili­arity with algori­thmic methods.
Task structure: Problems with clearly defined variables and rules favor algori­thmic approaches.
Motivation for accuracy: People are more likely to use algorithms when stakes are high.
Supportive enviro­nment: Tools like pen-an­d-p­aper, calcul­ators, or structured formats facilitate algorithm use.

Limita­tions in Human Use of Algorithms{{nl}}

Working memory constraints
Processing speed limitations
Suscep­tib­ility to fatigue or distraction
Tendency toward cognitive economy (favoring fast over correct answers)

Definition of Heuristics

Heuristics are mental shortcuts or informal rules of thumb that people use to make judgments, solve problems, and make decisions quickly and efficiently.
Unlike algori­thms, heuristics do not guarantee correct solutions, but they are cognit­ively economical and often sufficient in everyday contexts.
Origins and Develo­pment
The concept of heuristics became central to cognitive psychology in the 1970s through the work of Amos Tversky and Daniel Kahneman, who explored how people system­ati­cally deviate from rational judgment.
They identified heuristics as the cognitive tools that lead to biases in judgment and decision making.
Why We Use Heuristics
Cognitive economy: Heuristics reduce mental effort and processing time.
Limited inform­ation: People often make decisions with incomplete data.
Time pressure: Heuristics allow for quick decisions in urgent situations.
Uncert­ainty: Heuristics help navigate ambiguous or novel circumstances.
Adaptive value: In many situat­ions, heuristics lead to reasonably accurate outcomes.

Heuristics and Cognitive Biases

While heuristics are generally adaptive, they often lead to systematic errors or biases.
Examples of biases arising from heuristics include:
Confir­mation bias (favoring inform­ation that confirms prior beliefs)
Gambler’s fallacy (expecting outcomes to "­balance out")
Base rate neglect (ignoring statis­tical base rates in favor of vivid or specific details)

Relevance of Heuristics in Problem Solving

Heuristics often guide initial hypothesis formation and strategy selection in ill-de­fined problems.
In insigh­t-based or real-world problems, people frequently rely on intuitive rules rather than struct­ured, algori­thmic approa­ches.

Limita­tions of Heuristics

Can lead to biases and errors when used inappr­opr­iately
Overre­liance may prevent deeper analysis or re-eva­luation
Often contex­t-d­epe­ndent — what works well in one domain may fail in another
Difficult to detect or correct due to their uncons­cious, automatic nature

Theore­tical Frameworks of Heuristics

Bounded Ration­ality (Herbert Simon)
Humans are "­sat­isf­ice­rs" rather than optimizers — they seek satisf­actory solutions rather than perfect ones, especially when using heuris­tics.
Fast and Frugal Heuristics (Gerd Gigere­nzer)
Contrasts Tversky and Kahneman’s error-­focused view. Emphasizes that heuristics are often ecolog­ically rational and well-a­dapted to specific environments.
Argues that under certain condit­ions, heuristics outperform complex strategies.
Dual Process Theories
Heuristics are typically associated with System 1 thinking — fast, automatic, and intuitive — in contrast to System 2, which is slower and more analytical.
Tversky and Kahneman’s System 1/System 2 model is central to unders­tanding how heuristics are deployed in real-time decision making.

Functional Fixedness and Mental Set

Functional Fixedness
Mental Set
The cognitive bias that limits a person to using an object only in the way it is tradit­ionally used.
It prevents people from seeing altern­ative uses or functions for familiar tools and materials.
The tendency to approach problems using a strategy that has worked in the past, even when a newer, more efficient method is available.
First identified by Karl Duncker in the 1930s.
Theorist: Abraham Luchins (1942)
Classical Experi­ment: Duncker’s Candle Problem
Partic­ipants are given a candle, a box of tacks, and matches.
The task: Attach the candle to the wall so that it does not drip on the table.
Many fail to see the box as a platform rather than just a container, illust­rating functional fixedness.
Classical Experi­ment: Luchins’ Water Jar Problem
Partic­ipants are taught to use a complex formula (e.g., B - A - 2C) to solve several water jar volume problems.
Later, when a simpler method is possible, many still use the earlier complex strategy, showing a rigid mental set
Cognitive Explanation:
Arises from strong object­-fu­nction associ­ations stored in semantic memory.
Inhibits divergent thinking and insigh­t-based solutions.
Reflects how schema and experience can constrain perception of problem elements.
Cognitive Explanation:
Mental set reflects positive transfer that becomes maladaptive.
Reliance on familiar schemas blocks more efficient or creative solutions.
Involves automa­tiz­ation of procedures at the cost of flexib­ility.
Overcoming Functional Fixedness:
Reframing or recont­ext­ual­izing the problem.
Engaging in conceptual expansion (seeing familiar things in unfamiliar ways).
Encour­aging creativity and flexible thinking.
Reducing Mental Set Effects:
Training in flexible thinking and metacognition.
Awareness of cognitive biases.
Varied practice that discou­rages rigid rule-f­oll­owing.
 
Relation to Problem Space Theory (Newell & Simon):
Mental sets limit the explor­ation of altern­ative paths in the problem space.
They can cause a person to premat­urely settle into a fixed path or strategy.

Dual Process Theory of Thinking

Defini­tion:
Dual Process Theory suggests that human thinking operates through two distinct systems:
System 1: Fast, automatic, intuitive, and emotional.
System 2: Slow, delibe­rate, analyt­ical, and logical.

This theory explains why we sometimes rely on gut instincts, and other times on careful reasoning.
Key Features of the Two Systems
System 1
Operates automa­tically and quickly
Requires little or no effort
Based on heuristics (mental shortcuts)
Emotio­nally charged and context-dependent
Examples: Detecting hostility in a voice, driving a familiar route, solving 2 + 2
System 2
Allocates attention to effortful mental activities
Involves reasoning, logic, planning
Slower but more reliable
Used in unfamiliar or complex situations
Examples: Solving a math problem, evaluating an argument, planning a trip
Theorists
Daniel Kahneman and Amos Tversky popula­rized this model in the context of judgment and decision making.
Kahneman's book “Thinking, Fast and Slow” is a founda­tional text.
Applic­ations of Dual Process Theory
Explains biases and errors in decisi­on-­making (System 1 can be misleading)
Used in cognitive psycho­logy, behavioral economics, and education
Helps design better proble­m-s­olving strategies and interv­entions
Strengths
Explains both quick decisions and complex reasoning
Supported by research in neuros­cience and cognitive science
Helps account for cognitive biases and heuristics

Limita­tions
May oversi­mplify human thought into just two systems
Real thinking often involves intera­ction between the two
Boundaries between systems can blur
               
 

Comments

No comments yet. Add yours below!

Add a Comment

Your Comment

Please enter your name.

    Please enter your email address

      Please enter your Comment.

          Related Cheat Sheets

          6 Steps to FPS Cheat Sheet

          More Cheat Sheets by rentasticco

          Biological Bases of Behaviour Cheat Sheet
          Learning Cheat Sheet
          Attention Cheat Sheet