Final Exam Notes
Final Exam Notes
Cree
Semantic memory
● A type of long-term memory involving the capacity to recall words, concepts, or
numbers, which is essential for the use and understanding of language
● Tulving (1972) w as the first to define semantic memory
● This part illustrates how different sensory and functional aspects of objects (such as
action, touch, shape, color, sound) are integrated to form a cohesive representation
● Each of these aspects feeds into a central node, suggesting that our understanding of an
object is a composite of its different sensory inputs and associated actions
● Inanimate Objects: Patients could provide functional descriptions of these items (e.g.,
"Tent—temporary outhouse, living home.").
● Living Things: Patients' ability to describe these was impaired (e.g., "Parrot—don't
know.").
● This contrast highlights how damage to specific semantic subsystems affects the
ability to categorize and describe different types of objects, consistent with the
Sensory/Functional Theory
Brain scans
● There are two scans shown, likely indicating areas of the brain affected by damage that
leads to semantic deficits. These scans
support the idea that different brain
areas may be specialized for different
types of semantic processing
The term "double dissociation" comes into play when two patients have lesions in different parts
of the brain, and each patient's lesion leads to deficits in different cognitive functions. This helps
to demonstrate that the two cognitive functions are controlled by different regions of the brain
● This pattern suggests that the cognitive processes underlying the naming of animals and
the naming of certain foods are independent of each other and likely localized in different
brain areas
● This hypothesis suggests that different categories of objects are processed differently
in the brain, and certain neural circuits are specialized for processing certain types
of objects
○ The domain-specific hypothesis is important because it helps us understand how
information is organized in the brain
○ It suggests that there might be different regions of the brain that specialize in
processing different kinds of objects
Net Input
Activation
●
A little history…
➔ 1974: Back-Propagation was invented by a Harvard PhD student
◆ Neural network: feeds forward information
◆ Is an algorithm to test errors which will travel back from input nodes to output
nodes
● Applied to reduce cost function and errors
● The previous layer output is multiplied with weight which gives the
activation function (forward propogation)
● Output highest activation = suitable output marth for the corresponing
input
● BP is the tool to calculate the gradient of the lost function
○ Quick, easy, no extra functions
◆ Based on biological neural network
◆ Neurons = nodes
➔ 1980’s: rediscovered in the 1980’s by David Rumelhart and David Parker
➔ 1986: First popularized by Rumelhart, Williams, & Hinton (1986)
◆ Currently used to train neural nets in many different applications within
psychology and beyond (e.g., pre-cancerous cell detection, predict exchange rates,
adjust telescope optics, etc.)
Training set
Sentences were formed from
- Relative pronouns
- Who
- Nouns
- Boy, girl, cat, dog, boys,
girls, cats, dogs
- Proper nouns
- John
- Mary
- Verbs
- Chase
- Feed
- See
- Hear
- Walk
- Live
- Chases
- Feeds
- Sees
- Hears
- Walks
Sentences were created from these words folling a grammar check
- John feeds dogs
- Girls feed dogs
Conclusion
- The network developed hidden unit representations that can distinguish between the
occurrence of the same word (e.g., boy) in different grammatical positions in a sentence
(e.g., subject, object), and use that information to correctly predict the next word in the
sentence
- Elman interpreted this as evidence that neural networks could code for the complex
structural relationships inherent to language, and could learn this from mere exposure to
the language
- His model was just a toy example. It would take another 30 years for people to figure out
how to build more powerful systems that could do the same with real, natural language
- According to its creators; chatgpt can pass the bar exam, graduate record exam (GRE),
and score well on advanced placement tests
- What are the standards that constitute good LLM performance?
- Computer scientists: tends to output text that is grammatical, coherent, and natural
sounding
- Psychologists
➔ Therapeautic chatbots
➔ Tools that automate note taking and administrative tasks
➔ More intelligent training and intervention
● AI for providing new forms of data collection (e.g., monitor social media activity and
link to physiological metrics).
● AI as the participant.
● AI for writing code, including data analysis, data visualization, and more complex
projects.
● AI for writing and editing articles.
● AI for reproducing and extending analyses (e.g., repeat a meta-analysis that was
conducted 10 years ago including all the newly published data).
● AI for reviewing articles
Biases
● Because generative Large Language Models (LLMs) are trained on text produced by real
humans, they embody the same biases held by those humans, including gender, race, and
cultural biases.
● This has already led to discrimination due to the uncritical application of algorithms to
new tasks, such as filtering job applicant files.
● Significant, serious thought must be given to new applications in terms of what biases
could exist in the model due to the training set, and how things could go wrong if the
model is applied uncritically to new tasks.
● Most current AIs are created and owned by corporations, whose main motives are profit.
The motivations and goals of scientists, especially psychologists, are quite different,
involving stricter moral guidelines, including a duty to care for participants and clients
Week 11: Meta-Analysis
What is Meta Analysis? (The statistical synthesis of the results of different studies
addressing similar questions)
● First used in 1970
● Part of a broader field of research synthesis techniques that includes systematic reviews
Systematic reviews (a methodology for collecting, evaluating, and reporting of the data that go
into a meta-analysis (contrast with traditional narrative reviews). Consists of search, then
assessment
● Involves an a priori literature search strategy with replicable search terms and criteria for
selection of studies, reported in a protocol following standardized guidelines (e.g.,
PRISMA)
● Involves searches are conducted online using multiple scientific literature databases (e.g.,
Web of Science, Scopus, PsycInfo). Unpublished data may be included
● Involves search and screening should be done by at least 2 people
PRISMA
● Preferred Reporting Items for Systematic Reviews and Meta-Analyses
● Evidence-based minimum set of guidelines aimed at helping authors improve the
reporting of systematic reviews and meta-analyses
● PRISMA focuses on ways in which authors can ensure the transparent and complete
reporting of this type of research
● The guidelines include a 27-item checklist and a four-phase flow diagram to assist in the
systematic reporting of reviews and meta-analyses related to healthcare interventions,
though the principles can be applied more broadly to other fields as well. The aim is to
enhance the clarity, transparency, and rigor in the reporting of systematic reviews and
meta-analyses, facilitating the critical appraisal and replication of such studies
Meta Analysis Effect Sizes: We use effect sizes to summarize the findings that were reported
using different measures
● Once the systematic review of the literature is complete, we can start the meta-analysis
● First, we need to put the results of the different studies on the same scale
○ Analogy: cash gifts at wedding in different currencies - who gifted the most?
Meta Analysis Effect Sizes: Which effect size metric should you use?
● One common effect size metric is the standardized mean difference
● This is the difference between the means of two groups, divided by their pooled standard
deviation
● This is relatively easy to interpret: the effect size indicates how many standard deviations
the target group scores better (or worse) than the comparison group. 0.0 indicates no
difference
● Other alternatives include
○ Correlation coefficients
○ Log response ratio
○ Odds ratio
○ Hazard ratio
Meta Analysis Effect Sizes: What do you do if the results are reported in the opposite direction?
● Some studies may report results in the opposite direction of the results of the other
studies (e.g., most people study vitality, but some report fatigue).
● What should you do?
○ Reverse-score so that the higher scores indicate comparable responses, allowing
them to be meaningfully combined
● Overall, be clear about what you are measuring, and align this with the question(s) you
are asking
Meta Analysis Effect Sizes: Code for moderators at the same time
● Moderator: a third variable that influences the relationship between two other variables
(e.g., mental health may moderate the relationship between sleep quality and academic
achievement)
○ Examples include participant ages, gender, income, level of education, etc.
● As you are going through the studies, getting the effect sizes, note down characteristics of
the study that vary across studies and could potentially influence results
Aggregating Effect Sizes Across Studies (Basic premise: effect size from each study is
combined, weighted by sample size)
● Larger studies are assumed to be a better estimate of the “true effect”
● Why? Small studies are more susceptible to chance outcomes
● Therefore, we give larger studies more weight when we combine the results
Aggregating Effect Sizes Across Studies: Larger studies are a better estimate of the “true” effect
● If we know the statistical properties of an effect size metric, we can calculate the
sampling variance (the variance in that effect size if we were to do the same study many
times)
● This is what we use to weigh the larger studies more heavily
● The logic is that every study is assumed to estimate some “true” outcome. If the sample
size was large enough, we would get close to that true effect
Aggregating Effect Sizes Across Studies: Fixed vs. Random Effects Models
● If we assume that the only cause of variance in the effect sizes across studies is the
sampling variance, then we are using a
○ fixed effects model
● But in reality, there are likely to be more sources of variance than just the sampling
variance (e.g., location). If we factor these in, then we are using a
○ random effects model
● Fixed effects models are more like to generate false positives. They were popular when
meta-analysis first hit the scene, but random effects models are more popular today
Examning Heterogeneity
● Heterogeneity: the variability in the true effect across studies
● Imagine that we could run the same study multiple times, in the same way, with the same
experimental protocols, same participants, etc. Any differences in results would be due to
measurement error (difference between measured value of a quantity and its unknown
true value), and the studies would be homogeneous
● Heterogeneity denotes the variability in outcomes that goes beyond measurement error
alone
Examning Heterogeneity
● Heterogeneity: the way a study differs from the others, such as methodology,
characteristics of population, specifics of the intervention, etc. We lump these all together
into the heterogeneity term
● More heterogeneity means the mean effect size is less likely to reliably characterize
responses across studies, and therefore, may not be useful for providing sound
recommendations for practice or policy
● If there is a great deal of heterogeneity among studies, we interpret that to mean that the
true effect (the real response to the intervention) differs among studies
Indices of Heterogeneity
● Indices of heterogeneity help specify whether variation in the mean effect across studies
is reliably different from 0
● Example
○ Q-test: used to determine whether excess variation is present by change alone
○ I2: used to estimate the proportion of the total variation in true effect sizes that is
due to heterogeneity in the real responses across studies
■ 25 = low, 50 = medium, 75 = high heterogeneity
Study hetergoneity: I2
● I2 is a percentage that represents the proportion of observed variation that can be
attributed to the actual difference between studies (such as age of participants in different
studies) rather than within-study variance (differences between the individual participants
in a single study)
○ 25 = low; 50 = medium; 75 = high variance
● Not sensitive to the number of studies included