Show Menu
Cheatography

Learning in Psychology- Behaviourism, socio-cognitive, other theories

Defini­tions

Learning
A relatively permanent change in behaviour brought about by experience or practice.
 
"­Rel­atively perman­ent­"-- learning causes physical changes in the brain to record what has been learned.
Maturation
Change that is accomp­lished through biological growth, controlled by a genetic blueprint
Habitu­ation
A decline in the response to a stimulus once the stimulus has become familiar.
Dis-ha­bit­uation
An increase in respon­siv­eness when something novel is presented, following a series of presen­tations of something familiar.
Classical Condit­ioning
A form of learning in which one stimulus is paired with another so that the organism learns a relati­onship (implicit and involu­ntary) between the stimuli.
Uncond­itioned Stimulus
A stimulus that reliably triggers a particular response without prior training.
Uncond­itioned Response
A response elicited by an uncond­itioned stimulus without prior training.
Condit­ioned Stimulus
An initially neutral stimulus that comes to elicit a new response due to pairings with the uncond­itioned stimulus.
Condit­ioned Response
A response elicited by an initially neutral stimul­us—the condit­ioned stimulus (CS)—after it has been paired repeatedly with an uncond­itioned stimulus (US).
Second Order Condit­ioning
A form of learning in which a neutral stimulus is first made meaningful through classical condit­ioning. Then, that stimulus (the CS) is paired with a new, neutral stimulus until the new stimulus also elicits the condit­ioned response.
Extinction
The weakening of a learned response that is produced if a condit­ioned stimulus is now repeatedly presented without the uncond­itioned stimulus.
Sponta­neous Recovery
The reappe­arance of an exting­uished response after a period in which no further condit­ioning trials have been presented.
Stimulus Genera­liz­ation
The tendency for stimuli similar to those used during learning to elicit a reaction similar to the learned response.
Discri­min­ation
An aspect of learning in which the organism learns to respond differ­ently to stimuli that have been associated with a US (or reifor­cem­ent), and stimuli that have not.
Inhibitor
A stimulus signaling that an event is not coming, which elicits a response opposite to the one that the event usually elicits.
Blocking Effect
A result showing that an animal learns nothing about a stimulus if the stimulus provides no new inform­ation.
Compen­satory Response
A response that offsets the effects of the upcoming uncond­itioned stimulus.
Instru­mental Condit­ioning
A form of learning in which the partic­ipant receives a reinforcer only after performing the desired response, and thereby learns a relati­onship between the response and the reinfo­rcer. (Volun­tary, explicit)
Law of Effect
Thornd­ike’s theory that a response followed by a reward will be streng­thened, whereas a response followed by no reward (or by punish­ment) will be weakened.
Premack Principle
states that more probable behaviors will reinforce less probable behaviors.
Operant
In Skinner’s system, an instru­mental response that is defined by its effect (the way it operates) on the enviro­nment.
Reinforcer
A stimulus delivered after a response that makes the response more likely in the future.
Shaping
The process of eliciting a desired response by rewarding behaviors that are increa­singly similar to that response.
Behavi­oural Contrast
A response pattern in which an organism evalu- ates a reward relative to other avail- able rewards or those that have been available recently.
Partial Reinfo­rcement
A learning condition in which only some of the organism’s responses are reinforced
Schedule of Reinfo­rcement
The rules about how often and under what conditions a response will be reinfo­rced.
Ratio Schedule
A pattern of delivering reinfo­rce­ments only after a certain number of responses.
Interval Schedule
A pattern of delivering reinfo­rce­ments only after a certain amount of time has passed.
Latent Learning
Learning that occurs without a corres­ponding change in behavior.
Learned Helple­ssness
A condition of passivity apparently created by expo- sure to inesca­pable aversive events. This condition inhibits or prevents learning in later situations in which escape or avoidance is possible.
Observ­ational Learning
The process of watching how others behave and learning from their example.
Vicarious Condit­ioning
A form of learning in which the learner acquires a condit­ioned response merely by observing another partic­ipant being condit­ioned.
Mirror Neurons
Neurons that fire whenever an animal performs an action, such as stretching out its arm or reaching toward a target, and also whenever the animal watches another performing the same action.
Taste Aversion Learning
A form of learning in which an organism learns to avoid a taste after just one pairing of that taste with illness.
Prepared Learning
Learning that occurs without extensive training because of an evolved predis­pos­ition to the behavior.
Presyn­aptic Facili­tation
A process, documented in studies of Aplysia, that underlies many kinds of learning. It occurs when learning results in an increased release of neurot­ran­smitter into the synapse.
Long Term Potent­iation
A long-l­asting increase in a neuron’s response to specific inputs, caused by repeated stimul­ation.

Classical Condit­ioning

Ivan Pavlov: Won the nobel prize in 1904 for his work on digestive physio­logy: studying salivation in dogs
Discovery of Classical Condit­ioning
Noticed that dogs began to salivate not just at food (dry food = natural saliva­tion), but also at stimuli associated with food (like the dish, the person, or the lab). These neutral stimuli began to trigger salivation after being repeatedly paired with food.
Experiment Setup
Pavlov rang a bell (neutral stimulus) before giving food. After several repeti­tions, the bell alone triggered saliva­tion.
Uncond­itioned Stimulus (US): Food – naturally causes a reaction.
Uncond­itioned Response (UR): Salivation to food – natural, unlearned.
Condit­ioned Stimulus (CS): Bell – originally neutral, becomes meaningful through pairing.
Condit­ioned Response (CR): Salivation to the bell – learned response.
Acquis­ition of Condit­ioned Responses
At first, a Condit­ioned Stimulus (CS) (like a bell) does not cause a Condit­ioned Response (CR) (like saliva­tion). After repeated pairings with the Uncond­itioned Stimulus (US) (like food), the CS starts to trigger the CR. Learning is gradua­l—the strength of the CR builds up over time with more CS-US pairings.
Second­-Order Condit­ioning
Once a CS (e.g., light) has been paired with a US (e.g., food) to elicit a CR (saliv­ation), a new neutral stimulus (e.g., a bell) can be paired with the CS (light) to also trigger the CR—even without the US. This is called second­-order condit­ioning. Example: If the sight of a dentist causes fear due to painful experi­ences (US), then related cues (dentist's office, voice, etc.) can also trigger fear. It explains how fears or emotional responses spread through associ­ations.
Extinction
If the CS is presented without the US repeat­edly, the CR gradually weakens and disapp­ears. This is called extinc­tion. Example: If a bell is rung but no food follows, over time, the dog will stop salivating to the bell. Extinction is not the same as forget­ting: Forgetting is slow; extinction can happen in just a few trials. Evidence: After a delay (with no exposure), the CR can return (spont­aneous recovery).
Sponta­neous Recovery
After extinc­tion, if the animal is given a rest, the CS can again trigger the CR when presented. Shows that extinction doesn’t erase the original learni­ng—it just suppresses it. Sponta­neous recovery means the memory is still there; the animal is testing whether the CS is inform­ative again.
Recond­iti­oning
If the animal is condit­ioned again after extinc­tion, it learns much faster than the first time. Suggests that some memory of the original learning remains.
Real-Life Example: Exposure Therapy
Used to treat phobias and anxiety. The feared stimulus (CS) is presented without danger or trauma (US). Over time, anxiety (CR) decreases = extinc­tion. However, after therapy ends, anxiety can return = sponta­neous recovery. Not a failure of therapy, just a sign that more sessions are needed.
Genera­liz­ation
Defini­tion: The tendency of a learned response (CR) to occur in the presence of stimuli that are similar, but not identical, to the original condit­ioned stimulus (CS).
Example: A dog trained to salivate at a specific tone will also salivate (less strongly) to other, similar tones.
Genera­liz­ation Gradient: The more different a new stimulus is from the original CS, the weaker the condit­ioned response becomes.
Discri­min­ation
Defini­tion: The ability to distin­guish between different stimuli, responding only to the CS+ (which is followed by the US) and not to similar but non-pr­edi­ctive stimuli (CS–).
Example: If a red light (CS+) signals a boat horn (US), and an orange light (CS–) never does, a person will eventually tense up only to the red light.
CS– Role: It signals the absence of the US, becoming an inhibitor that reduces the likelihood of the CR.
CS as a “Signal”
The CS works best when it predicts the arrival of the US.
Timing Matters: Forward pairing (CS before US, short delay): Most effective. Simult­aneous pairing (CS and US at the same time): Less effective. Backward pairing (US before CS): Least effective. Analogy: Like a caution sign before a dangerous curve: Just before the curve = effective (CS predicts US). Too early = ineffe­ctive (CS too far ahead of US). During or after the curve = useless or confusing (simul­taneous or backward pairing).
Contin­gency vs. Contiguity
Contiguity means the CS (Condi­tioned Stimulus) and US (Uncon­dit­ioned Stimulus) occur close in time. Contin­gency means the CS predicts the likelihood of the US. Key Insight: Learning doesn’t happen just because two things happen close together (conti­guity); instead, learning depends on whether the CS provides useful inform­ation about the US (conti­nge­ncy). 🐶 Example: The dog hears a metronome and gets food. Many other things (light, noise) are also present. But only the metronome reliably signals the arrival of food. That’s contin­gency.
The Role of Inform­ation
Animals (and humans) learn only from stimuli that give reliable info about what’s going to happen. If a stimulus is always present, regardless of whether the US comes or not (like light fixtures), it gives no predictive value and won’t be learned as a signal.
Experiment on Rats – The Role of Predictive Value
Group A: Shock sometimes follows the bell, but it also happens just as often without the bell → no contin­gency, no learning. Group B: Shock more likely after the bell than without → some contin­gency, learning occurs. Conclu­sion: Even an imperfect predictor (40% chance) can cause condit­ioning if it increases the likelihood of a US compared to baseline.
The Absence of Contin­gency
When tones and shocks are randomly paired, there’s no way to know when a shock is coming → no condit­ioning happens. If shocks only follow tones, even incons­ist­ently (e.g. 50% of the time), animals learn because the tone predicts something. Key Concept: Unpred­ict­ability leads to chronic stress. When there’s a danger signal (e.g. tone), there’s also a sense of safety when it’s absent. Random shocks = constant anxiety.
Rescor­la-­Wagner Model
Learning happens when there’s a surprise, and it stops when things become predic­table.
Core Idea
Your brain is constantly trying to predict what will happen. If something happens that’s unexpe­cted, your brain says: "­Whoa! I didn’t see that coming — I need to learn from this!"
How It Works
Let’s say the brain keeps a “score” of how much it expects the US (e.g., food or shock) after a CS (like a tone). 🧾 The Learning Formula: Change in learning = How surprising the US is = (What actually happened) – (What was expected) Or: ΔV = λ – V Where: ΔV = change in strength of learning λ = the actual outcome (Was there food/s­hock? How strong?) V = what was expected (How much did the animal already think the food/shock was coming?)
Key Points
Learning = Surprise No surprise = no learning. Prediction gets updated each time based on error (diffe­rence between expected and actual outcome). Eventu­ally, when predic­tions are perfect, learning stops. If a new CS (like a light) adds no new inform­ation, it won’t be learned — this explains blocking (an advanced concept, but tied to this model).
CR and UR are NOT always the same
UR (Uncon­dit­ioned Response) happens naturally after the US (Uncon­dit­ioned Stimulus). CR (Condi­tioned Response) is a learned reaction to the CS (Condi­tioned Stimulus). They might look different even though they're both “respo­nses.”
Example:
Rat + shock (US) → jumps and squeals (UR) Flashing light (CS) → freezes and heart slows (CR = preparing for shock)
CR = “Get ready!” response Not a direct copy of the UR—it's an antici­patory adjust­ment.
CR helps the body prepare
Animals aren’t reacting randomly — the CR prepares them for what's coming.
Example: Tone → signals food is coming → dog moistens mouth Light → signals shock → animal freezes, alert 📌 This prepar­ation makes the animal more efficient or safe.
Condit­ioning and Drug Tolerance
Repeated drug use (like heroin) leads to tolerance: you need more to feel the same effect. Why? Because the body learns to compensate in advance. How? US = heroin UR = drug’s biological effects (pain relief, dry mouth, good mood) CS = sight of needle, drug enviro­nment CR = opposite of the drug’s effects (more pain sensit­ivity, bad mood, wet mouth) 📌 CR is a compen­satory response = the body is trying to stay in balance (homeo­sta­sis).
Drug Craving = CR with no US
If a user sees the CS (needle, place, routine) but no drug arrives, the CR still happens. So they feel pain, depres­sion, cravings — the opposite of what the drug would’ve done. 📌 Craving = your body bracing for a drug that doesn’t come.

Insight Learning

What is Insight Learning?
Insight learning is a type of learning that happens suddenly, through unders­tanding relati­onships between different parts of a problem, rather than through trial-­and­-error.
 
It involves a sudden realiz­ati­on—an “Aha!” moment­—where the solution just clicks.
 
Insight is not based on condit­ioning or reinfo­rce­ment, but on cognitive restru­cturing of the problem.
Background & Theorist: Wolfgang Köhler
 
Köhler was a Gestalt psycho­logist who studied proble­m-s­olving in chimpa­nzees.
 
Gestalt psychology emphasizes holistic proces­sin­g—how we perceive whole patterns, not just bits and pieces.
Köhler’s Famous Experi­ments with Chimps
🐒 Example 1: Sultan and the Stick Setup: A chimpanzee named Sultan was placed in a cage with a banana just out of reach, and sticks nearby. Process: Sultan tried reaching it unsucc­ess­fully, then stopped and seemed to think. Insight: Suddenly, Sultan used one stick to pull another closer and joined them to reach the banana. 👉 He did not arrive at this through gradual trial-­and­-er­ror—it came suddenly.
🐒Example 2: Box Stacking Chimps were given boxes and a hanging banana. They stacked the boxes and climbed them to get the banana, showing unders­tanding of spatial relati­ons­hips.
Key Charac­ter­istics of Insight Learning
Charac­ter­istic
Descri­ption
Suddenness
The solution appears all at once (the “aha!” experi­ence).
Unders­tanding
Involves grasping the structure of the problem, not random attempts.
No trial-­and­-error
Unlike condit­ioning, it doesn’t rely on repeated errors or reinfo­rce­ment.
Transf­era­bility
The solution or principle can often be applied to new but similar problems.
Requires mental reorga­niz­ation
Learner reinte­rprets the problem and mentally restru­ctures the elements
Insight Learning vs. Trial-­and­-Error Learning
Insight Learning
Trial-­and­-Error Learning
Cognitive and sudden
Behavioral and gradual
Based on perception and problem analysis
Based on repeated attempts and failures
May take longer to reach, but solution is quick
Gradual improv­ement over time
Solving a riddle by rethinking it
Trying keys one by one to open a lock
📌 Implic­ations of Insight Learning
Shows the Role of Cognition: Learning isn’t always about reinfo­rce­men­t—t­hinking matters
Applies to Humans & Animals: Though common in humans (espec­ially in proble­m-s­olv­ing), it has been shown in chimpa­nzees, birds, and other species.
Relevance in Education: Encourages the design of learning enviro­nments that foster critical thinking rather than rote memori­zation. Helps explain creative proble­m-s­olving in real-life situat­ions.
Real-Life Examples: Figuring out a tricky riddle after staring at it for a while. A child suddenly realizing how to tie shoelaces after watching but not previously succee­ding. An inventor seeing a solution to a problem after stepping away from it and then suddenly “seeing” the answer.
 

Operan­t/I­nst­rum­ental Condit­ioning

What is Instru­mental Condit­ioning?
Also known as Operant Condit­ioning. Involves learning voluntary behaviors (as opposed to automatic reflexes). Behavior is initiated by the organism, not triggered by an external stimulus. The outcome of the behavior (its conseq­uence) shapes future behavior.
Classical vs. Operant Condit­ioning
Classical: Involu­nta­ry/­ref­lexive responses (e.g., saliva­tion). Instru­mental: Voluntary actions (e.g., pressing a lever to get food).
Thorndike and the Law of Effect
Puzzle Box Experi­ment:
Hungry cats placed in a box with a mechanism (like a loop or lever) to escape and access food. First attempts: random behaviors (biting, scratc­hing). Eventu­ally, by trial and error, they hit the correct action. With repeti­tion, escape time gradually decrea­sed­—in­dic­ating learning.
Insight or Gradual Learning?
Thorndike found no sudden "­Aha­!" moment. Learning curve was gradual, not abrupt → suggests no reasoning, just reinfo­rce­ment.
The Law of Effect:
Behavior followed by a reward → streng­thened. Behavior followed by no reward or punishment → weakened. The animal doesn’t need insight or unders­tan­din­g—just conseq­uences. 📌 Learning = responses are "­stamped in" (if rewarded) or "­stamped out" (if not). 🔁 Parallel to Natural Selection: Like evolution: successful behaviors "­sur­viv­e", and useless ones fade. No conscious direct­ion­—just selection based on outcomes.
Skinner and Operant Behavior
Key Ideas: Distin­guished operant from classical condit­ioning: Classical: Response is elicited by a stimulus. Operant: Response is emitted volunt­arily by the organism. Called these voluntary responses operants because they operate on the enviro­nment. 📍 Core principle: Behavior + Positive Conseq­uence = More Likely in Future Behavior + Negative Conseq­uence = Less Likely
The Skinner Box: A controlled chamber where animals (rats, pigeons) could perform behaviors like pressing a lever or pecking a key for food. Allowed for rapid, repeated trials. Measured response rate = # of behaviors per unit of time. ✅ Advantage: More efficient than Thornd­ike's puzzle box (didn't need to reset after each trial).
Differ­ences Between Classical and Instru­mental Condit­ioning
Classical Condit­ioning: Learning about the relati­onship between two stimuli (e.g., bell and food). The response is automatic or reflexive (UR). Instru­mental Condit­ioning (Operant Condit­ion­ing): Learning about the relati­onship between a response and its conseq­uence (reinf­orcer or punish­ment). The response is voluntary. Despite differ­ences, both involve learning relati­onships among events and share key phenomena (like extinc­tion, genera­liz­ation, and discri­min­ation).
Learning Trials and Extinction
In classical condit­ioning: CS followed by US leads to learning. In instru­mental condit­ioning: Response followed by reinforcer leads to learning. Extinction happens when reinfo­rcement stops. The behavior gradually weakens or disapp­ears.
Genera­liz­ation and Discri­min­ation Genera­liz­ation: After learning a response to one stimulus (S+), animals often respond similarly to similar stimuli. The further the test stimulus is from the original S+, the weaker the response (seen in pigeons trained with light colors). Discri­min­ation: Animals learn to distin­guish between stimuli that signal different outcomes: S+ (positive discri­min­ative stimulus): Signals reinfo­rce­ment. S– (negative discri­min­ative stimulus): Signals no reinfo­rce­ment. Example: A child behaves better when parents are around (S+) than when they’re not (S–).
Complex Discri­min­ations Animals (like pigeons) can make surpri­singly complex discri­min­ations: Water vs. non-water pictures Trees vs. non-trees Recogn­izing individual humans from varied angles Shows learning goes beyond simple sensory cues and includes abstract catego­ries.
Shaping (Succe­ssive Approx­ima­tions) Used to teach complex or unlikely behaviors. Reinfo­rcement is given step-b­y-step as the animal’s behavior gradually approx­imates the desired action. Example: To teach a rat to press a high lever: Reinforce being near the lever → facing it → raising head → touching lever → pressing lever.
What is a Reinfo­rcer? Primary Reinfo­rcers: Naturally rewarding (e.g., food, water). Social Reinfo­rcers: Praise, smiles, etc. Condit­ioned Reinfo­rcers: Gain value by being associated with primary reinfo­rcers (e.g., money). Some reinfo­rcers are inform­ational or experi­ential (e.g., watching a toy train, using a wheel).
Behavioral Contrast The effect­iveness of a reinforcer depends on context and past experi­ence. A reward may seem large or small depending on what the subject was used to before. Example: 16 food pellets feel small after 60 but generous after 4.
Intrinsic Motivation and Overju­sti­fic­ation Effect Overju­sti­fic­ation Effect: External rewards can reduce intrinsic interest in an activity. Example: Children who initially liked drawing became less interested after being rewarded with “Good Player” certif­icates and then having those rewards removed.
Schedules of Reinfo­rcement
Partial Reinfo­rce­ment: Behavior is reinforced only sometimes, not every time. Effect: Leads to greater resistance to extinction (we keep trying even if not always rewarded).
Types of Reinfo­rcement Schedules Ratio Schedules (Based on number of responses) Fixed Ratio (FR): Reinfo­rcement after a set number of responses. E.g., FR 2 → reward after every 2 responses. Variable Ratio (VR): Reinfo­rcement after a varying number of responses (avera­ge-­based). E.g., VR 10 → reward on average after 10 responses (might be 5, then 15, etc.). Common in gambling (e.g., slot machines). Interval Schedules (Based on passage of time) Fixed Interval (FI): First response after a fixed time is rewarded. E.g., FI 3 minutes → reward given after 3 minutes if response made. Variable Interval (VI): Time interval changes; average time determines reward schedule. E.g., VI 8 minutes → reward on average after 8 minutes.
Contin­gency in Instru­mental Condit­ioning
Contin­gency: Behavior must predict the reward—not just follow it. Similar to classical condit­ioning: Predic­tion, not just pairing, is key. The likelihood of reward must be greater with the behavior than without it. 🔧 Control Matters: Organisms like having control over outcomes. When indivi­duals can predict and influence rewards, learning is enhanced.
Experi­ment: Infants and Mobiles
Group 1: Infants could control the mobile by moving their heads. Result: Enjoyed and engaged with the mobile. Group 2: Mobile moved indepe­nde­ntly, not due to the infant’s action. Result: Infants lost interest. Key Point: Even 2-mont­h-old infants prefer control and enjoy mastery.
Learned Helple­ssness
The Dog Study: Group A: Dogs could stop shocks by pressing a panel. Group B: Received same shocks, but had no contro­l—s­hocks were inesca­pable. Later task: Jump a barrier to avoid shock. Group A: Learned and escaped quickly. Group B: Became passive, didn’t try to escape­—even though it was possible.
Learned Helple­ssness: When previous lack of control leads to a belief that future attempts are useless. Leads to passivity, even when escape or success is possible.

Latent Learning

Edward Tolman: Believed learning is more than behavior change­—it's acquiring knowledge.
Latent Learning:
Learning that occurs without any obvious reinfo­rcement and does not immedi­ately manifest in behavior. The learned knowledge becomes apparent only when there is motivation to demons­trate it.
Classic Experiment (Tolman & Honzik, 1930):
Rats explored a maze for 10 days with no reward → no visible change in behavior. On Day 11, food was introduced at the goal box → rats quickly and accurately ran to the food. This showed they had formed a mental map of the maze during unrewarded explor­ation.
Mental Maps (Cognitive Maps):
Internal repres­ent­ations of the enviro­nment. Allow organisms to navigate spaces effici­ently even without direct reinfo­rce­ment.
Implic­ations
Learning ≠ Immediate Behavior Change: Just because behavior hasn’t changed doesn’t mean learning hasn’t happened. Supports Cognitive Perspe­ctive: Emphasizes the role of internal cognitive processes, not just stimul­us-­res­ponse links. Challenge to Behavi­orism: Opposes the strict behavi­orist view (e.g., Thornd­ike’s law of effect) that learning only occurs via reinfo­rce­ment. Practical Relevance: Students may learn a lot during lectures without immedi­ately demons­trating it. Skills and knowledge can emerge when they become relevant or useful. Used by Many Species: Not limited to humans­—many animals develop mental maps for foraging, naviga­tion, etc.

Observ­ational Learning

What is Observ­ational Learning?
Learning by watching others and imitating their behavior.
Also called social learning or vicarious learning.
No direct experience or reinfo­rcement needed­—just observ­ation.
Once thought to be uniquely human, but now observed in many animals too. 📚 Examples: Monkeys: Learn fear by watching another monkey react fearfully. Pigeons: Imitate behaviors like pecking or stepping to get rewards after watching others.
Mirror Neurons
Special neurons in the frontal lobe near the motor cortex. Fire when you perform an action and when you see someone else perform the same action. Help with unders­tanding others’ actions and imitating them. Found in many species, including humans.
Human Imitation Starts Early
Infants imitate facial expres­sions within the first month of life. Later, they mimic a wide range of behaviors. Two types of imitation: Mimicry: Copying the exact behavior. Modeling: Learning general rules or what behavior is "­oka­y" in a setting.
Deferred imitation: means imitating an action after a delay, not right after seeing it. The behavior is observed first, then reproduced later, sometimes hours, days, or even weeks afterward.
Bandura's Bobo Doll Experiment
Albert Bandura wanted to test whether children learn aggressive behavior by observing adults.
Experiment Setup
Partic­ipants: Preschool children (around 3–6 years old) Groups: Aggressive model: Kids watched an adult physically and verbally attack a Bobo doll (e.g., hitting, kicking, saying "­pow­!"). Non-ag­gre­ssive model: Kids watched an adult play quietly and nicely with toys. Control group: Kids saw no model.
Results
After watching the adult, kids were taken to a room with toys including a Bobo doll. Children who saw the aggressive model were more likely to imitate the aggression — even using the same actions and words. Some children went beyond imitation, showing new aggressive behaviors.
Key Findings
Children learn social behavior like aggression through observ­ation, not just direct reinfo­rce­ment. Modeling matters: kids imitate what they see, especially if the model is powerful or similar to them. This learning can be immediate or delayed (deferred imitat­ion). Boys showed more physical aggression than girls, but both imitated the behavior.
4 Steps of Observ­ational Learning
Attention: You have to notice the behaviour
Retention: You have to remember the behaviour
Reprod­uction: You must be able to replicate it
Motiva­tion: You must want to do it
Vicarious reinfo­rce­ment: Vicarious reinfo­rcement is when we learn by watching someone else get rewarded for a behavior — and then we're more likely to do that behavior ourselves.
Charac­ter­istics of the Model That Affect Learning:
Perceived Similarity People are more likely to imitate models who are similar to them (in age, gender, interests, etc.).
Perceived Competence If the model appears skilled or knowle­dge­able, observers are more likely to imitate them.
Status and Prestige Models with high social status (celeb­rities, teachers, respected peers) are more influe­ntial.
Warmth and Nurturance Models who are kind, friendly, and caring tend to be imitated more often.
Power or Authority Models who hold authority or power (like parents or police officers) can strongly influence behavior.
Consis­tency of Behavior Consistent behavior across situations makes a model more trustw­orthy and worth copying.
Reinfo­rcement or Punishment Observed If the model is rewarded, the observer is more likely to imitate. If the model is punished, the behavior is less likely to be copied.
 

Comments

No comments yet. Add yours below!

Add a Comment

Your Comment

Please enter your name.

    Please enter your email address

      Please enter your Comment.

          Related Cheat Sheets

          Spanish Vocabulary Cheat Sheet
          Italian Vocabulary Cheat Sheet
          Studying Cheat Sheet

          More Cheat Sheets by rentasticco

          Biological Bases of Behaviour Cheat Sheet