Internals Answers
Internals Answers
What is ANOVA
ANOVA is a statistical technique that assesses potential differences in a scale-level
dependent variable by a nominal-level variable having 2 or more categories. For example, an
ANOVA can examine potential differences in IQ scores by Country (US vs. Canada vs. Italy
vs. Spain). The ANOVA, developed by Ronald Fisher in 1918, extends the t and the z test
which have the problem of only allowing the nominal level variable to have two categories.
This test is also called the Fisher analysis of variance.
Definition
A study that compares patients who have a disease or outcome of interest (cases) with
patients who do not have the disease or outcome (controls), and looks back retrospectively to
compare how frequently the exposure to a risk factor is present in each group to determine
the relationship between the risk factor and the disease.
Case control studies are observational because no intervention is attempted and no attempt
is made to alter the course of the disease. The goal is to retrospectively determine the
exposure to the risk factor of interest from each of the two groups of individuals: cases and
controls. These studies are designed to estimate odds.
Case control studies are also known as "retrospective studies" and "case-referent studies."
Advantages
Good for studying rare conditions or diseases
Less time needed to conduct the study because the condition or disease has already
occurred
Lets you simultaneously look at multiple risk factors
Useful as initial studies to establish an association
Can answer questions that could not be answered through other study designs
Disadvantages
Retrospective studies have more problems with data quality because they rely on
memory and people with a condition will be more motivated to recall risk factors
(also called recall bias).
Not good for evaluating diagnostic tests because it’s already clear that the cases
have the condition and the controls do not
It can be difficult to find a suitable control group
Design pitfalls to look out for
Care should be taken to avoid confounding, which arises when an exposure and an outcome
are both strongly associated with a third variable. Controls should be subjects who might
have been cases in the study but are selected independent of the exposure. Cases and controls
should also not be "over-matched."
Is the control group appropriate for the population? Does the study use matching or pairing
appropriately to avoid the effects of a confounding variable? Does it use appropriate
inclusion and exclusion criteria?
Introduction
The first steps in learning to understand and appreciate evidence-based medicine are daunting
to say the least, especially when confronted with the myriad of statistics in any paper. This
short tutorial aims to introduce healthcare students to the interpretation of some of the most
commonly used statistics for reporting the results of medical research.
The scenario for this tutorial is centred around the diagram below, which outlines a fictional
parallel two arm randomised controlled trial of a new cholesterol lowering medication against
a placebo.
P values
P < 0.05 indicates a statistically significant difference between groups. P>0.05 indicates there
is not a statistically significant difference between groups.
Concept check 3
So if the trial comparing SuperStatin to placebo stated OR 0.5 95%CI 0.4-0.6 p<0.01
What would it mean?
A) The odds of death in the SuperStatin arm are 50% less than in the placebo arm with the
true population effect between 60% and 40%. This result was statistically significant.
B) The odds of death in the SuperStatin arm are 50% less than in the placebo arm with the
true population effect between 60% and 40%. This result was not statistically significant.
C) The odds of death in the SuperStatin arm are 50% less than in the placebo arm with the
true population effect between 60% and 40%. This result was equivocal.
Summary
This is a very basic introduction to interpreting odds ratios, confidence intervals and p values
only and should help healthcare students begin to make sense of published research, which
can initially be a daunting prospect. However it should be stressed that any results are only
valid if the study was well designed and conducted, which highlights the importance of
critical appraisal as a key feature of evidence based medicine.
2. Logistic Regression
Logistic regression is used to find the probability of event=Success and event=Failure. We
should use logistic regression when the dependent variable is binary (0/ 1, True/ False, Yes/
No) in nature. Here the value of Y ranges from 0 to 1 and it can represented by following
equation.
3. Polynomial Regression
A regression equation is a polynomial regression equation if the power of independent
variable is more than 1.
4.Stepwise Regression
This form of regression is used when we deal with multiple independent variables. In this
technique, the selection of independent variables is done with the help of an automatic
process, which involves no human intervention.
This feat is achieved by observing statistical values like R-square, t-stats and AIC metric to
discern significant variables. Stepwise regression basically fits the regression model by
adding/dropping co-variates one at a time based on a specified criterion. Some of the most
commonly used Stepwise regression methods are listed below:
Standard stepwise regression does two things. It adds and removes predictors as
needed for each step.
Forward selection starts with most significant predictor in the model and adds variable
for each step.
Backward elimination starts with all predictors in the model and removes the least
significant variable for each step.
The aim of this modeling technique is to maximize the prediction power with minimum
number of predictor variables. It is one of the method to handle higher dimensionality of data
set.
5. Ridge Regression
Ridge Regression is a technique used when the data suffers from multicollinearity
( independent variables are highly correlated). In multicollinearity, even though the least
squares estimates (OLS) are unbiased, their variances are large which deviates the observed
value far from the true value. By adding a degree of bias to the regression estimates, ridge
regression reduces the standard errors.
select the right regression model?
Life is usually simple, when you know only one or two techniques. One of the training
institutes I know of tells their students – if the outcome is continuous – apply linear
regression. If it is binary – use logistic regression!
NMHS and the sampling for the study, how to calculate prevalence,
DEGREES OF FREEDOM
Definition: The Degrees of Freedom refers to the number of values involved in the
calculations that have the freedom to vary. In other words, the degrees of freedom, in general,
can be defined as the total number of observations minus the number of independent
constraints imposed on the observations.
The degrees of freedom are calculated for the following statistical tests to check their
validity:
1. t-Distribution
2. F- Distribution
3. Chi-Square Distribution
These tests are usually done to compare the observed data with the data that is expected to be
obtained with a specific hypothesis.
Degrees of Freedom is usually denoted by a Greek symbol ν (mu) and is commonly
abbreviated as, df. The statistical formula to compute the value of degrees of freedom is
quite simple and is equal to the number of values in the data set minus one. Symbolically:
df= n-1
Where n is the number of values in the data set or the sample size. The concept of df can be
further understood through an illustration given below:
Suppose there is a data set X that includes the values: 10,20,30,40. First of all, we will
calculate the mean of these values, which is equal to:
(10+20+30+40) /4 = 25.
Once the mean is calculated, apply the formula of degrees of freedom. As the number of
values in the data set or sample size is 4, so,
df = 4-1=3.
Thus, this shows that there are three values in the data set that have the freedom to vary as
long as the mean is 25.
How do you find the degrees of freedom for a paired sample t test?
The most commonly encountered equation to determine degrees of freedom in statistics
is df = N-1
What is the calculation of the degrees of freedom for the two independent sample t test?
In the case of a t-test, there are two samples, so the degrees of freedom are N1 + N2 – 2 = df.
Once you determine the significance level (first row) and the degrees of freedom (first
column), the intersection of the two in the chart is the critical value for your particular study.
dF for ANOVA
One way ANOVA
DF1= k(no of groups) -1
Two way Anova 2*3 design
Df1= (2-1) *(3-1)
For the case of within subjects-designs, things can become a bit more complicated. The
following paragraphs are work in progress. The calculation of df2 for a repeated measures
ANOVA with one within-subjects factor is as follows: df2 = df_total – df_subjects –
df_factor, where df_total = number of observations (across all levels of the within-subjects
factor, n) – 1, df_subjects = number of participants (N) – 1, and df_factor = number of levels
(k) – 1. Basically, the take home message for repeated measures ANOVA is that you lose one
additional degree of freedom for the subjects
CHI SQUARE
The Chi-square statistic is a non-parametric (distribution free) tool designed to analyze
group differences when the dependent variable is measured at a nominal level. ... The
Cramer's V is the most common strength test used to test the data when a significant Chi-
square result has been obtained.
O = the Observed (actual) value
E = the Expected value
T-tests are easier to conduct, so why not conduct a t-test for the possible
interactions in the experiment? A Type I error is the answer because the more
hypothesis tests you use the more you risk making a type I error and the less
power a test has. There is no disputing the t-test changed statistics with its
ability to find significance with a small sample, but as previously mentioned the
ANOVA allowed for testing more than 2 means. ANOVA’s are used a lot
professionally when testing pharmaceuticals and therapies.
The ANOVA is an important test because it enables us to see for example how
effective two different types of treatment are and how durable they are.
Effectively a ANOVA can tell us how well a treatment work, how long it lasts
and how budget friendly it will be an example being intensive early behavioural
intervention (EIBI) for autistic children which lasts a long time with a lot hour,
has amazing results but costs a lot of money. The ANOVA is able to tell us if
another therapy can do the same task in shorter amount of time and therefor
costing less and making the treatment more accessible. Conducting this test
would also help establish concurrent validity for the therapy against EIBI. The F-
ratio tells the researcher how big of a difference there is between the conditions
and the effect is more than just chance. ANOVA test assumes three things:
The population sample must be normal
The observations must be independent in each sample
The population the samples are selected from have equal variance
a.k.a. homogeneity of variance.
These requirements are the same for a paired and a repeated measures t-test and
these measured are solved in the same way for the t-test and the ANOVA. The
population sample is assumed to be normal anyway, the independent samples is
achieved with the design of the experiment, if the variance is not correct then
normally more data (participants) is needed in the experiment.
In conclusion it is necessary to use the ANOVA when the design of a study has
more than 2 condition to compare. The t-test is simple and less daunting
especially when you see a 2x4x5 factorial ANOVA is needed, but the risk of
committing a type I error is not worth it. The time you spent conducting the
experiment only to have it declared obsolete because the right statistical test
wasn’t conducted would be a waste of time and resources, statistical tests should
be used correctly for this reason.
standard error
The following table lists the nonparametric tests and their parametric alternatives.
N ON P A RA M ETRIC TES T P A RA M ETRIC A LTERN A TIV E
1-sample sign test One-sample Z-test, One sample t-test
1-sample Wilcoxon Signed Rank test One sample Z-test, One sample t-test
Friedman test Two-way ANOVA
Kruskal-Wallis test One-way ANOVA
Mann-Whitney test Independent samples t-test
Mood’s Median test One-way ANOVA
Spearman Rank Correlation Correlation Coeff
Gold standard tests mean that diseases and conditions can be correctly classified.
Examples
Breast CancerMammogramTissue Biopsy
D IS EAS E/CO N D ITIO N IN ITIA L TES T G O LD S TA N DA R D
Rabies (in animals) ? DFA test
Tuberculosis Tuberculin skin or blood test Lowenstein-Jensen culture
Colorectal cancer Blood in stool sample Colonoscopy
Diabetes Fasting plasma glucose Oral GTT
Assessing new tests
New diagnostic tests are compared against the gold standard. Sensitivity and specificity for
these new tests are usually estimated from comparing them to the gold standard. This system
is not perfect: no test (even if the best available) is perfect and has some bias (error) attached
to it. For example, biopsies vary wildly in accuracy depending on the specific cancer type.
What this means is you’re essentially comparing any new test to a standard that has some
error attached it it — which means your new test will also have error. This can be avoided by
using a third, “resolver” test, followed by re-testing some people whose results are the same
(either positive or negative) on both the new test and the resolver test.
CRITICAL VALUE,
What is a critical value in statistics?
In hypothesis testing, a critical value is a point on the test distribution that is compared to the
test statistic to determine whether to reject the null hypothesis. If the absolute value of your
test statistic is greater than the critical value, you can declare statistical significance and
reject the null hypothesis.
When the sampling distribution of the statistic is normal or nearly normal, the critical value
can be expressed as a t score or as a z-score. To find the critical value, follow these steps.
Compute alpha (α): α = 1 - (confidence level / 100)
Find the critical probability (p*): p* = 1 - α/2
To express the critical value as a z-score, find the z-score having a cumulative
probability equal to the critical probability (p*).
To express the critical value as a t statistic, follow these steps.
Find the degrees of freedom (df). Often, df is equal to the sample size minus
one.
The critical t statistic (t*) is the t statistic having degrees of freedom equal to
df and a cumulative probability equal to the critical probability (p*).
Should you express the critical value as a t statistic or as a z-score? There are several ways to
answer this question. As a practical matter, when the sample size is large (greater than 40), it
doesn't make much difference. Both approaches yield similar results. Strictly speaking, when
the population standard deviation is unknown or when the sample size is small, the t statistic
is preferred. Nevertheless, many introductory texts and the Advanced Placement Statistics
Exam use the z-score exclusively.
MEASURES OF DISPERSION
What do measures of dispersion show?
As the name suggests, the measure of dispersion shows the scatterings of the data. It tells
the variation of the data from one another and gives a clear idea about the distribution of the
data. The measure of dispersion shows the homogeneity or the heterogeneity of the
distribution of the observations
STANDARD DEVIATION
The standard deviation is simply the square root of the variance. This makes the standard
deviations of the two quiz distributions 1.225 and 2.588. The standard deviation is an
especially useful measure of variability when the distribution is normal or approximately
normal (see Chapter on Normal Distributions) because the proportion of the distribution
within a given number of standard deviations from the mean can be calculated. For example,
68% of the distribution is within one standard deviation of the mean and approximately 95%
of the distribution is within two standard deviations of the mean. Therefore, if you had a
normal distribution with a mean of 50 and a standard deviation of 10, then 68% of the
distribution would be between 50 - 10 = 40 and 50 +10 =60. Similarly, about 95% of the
distribution would be between 50 - 2 x 10 = 30 and 50 + 2 x 10 = 70. The symbol for the
population standard deviation is σ; the symbol for an estimate computed in a sample is s.
Figure 2 shows two normal distributions. The red distribution has a mean of 40 and a
standard deviation of 5; the blue distribution has a mean of 60 and a standard deviation of 10.
For the red distribution, 68% of the distribution is between 35 and 45; for the blue
distribution, 68% is between 50 and 70.
Level Of Significance
The level of significance is defined as the probability of rejecting a null hypothesis by the test
when it is really true, which is denoted as α. That is, P (Type I error) = α.
Confidence level:
Confidence level refers to the possibility of a parameter that lies within a specified range of
values, which is denoted as c. Moreover, the confidence level is connected with the level of
significance. The relationship between level of significance and the confidence level is
c=1−α.
The common level of significance and the corresponding confidence level are given below:
Rejection region:
The rejection region is the values of test statistic for which the null hypothesis is rejected.
The set of all possible values for which the null hypothesis is not rejected is called the
rejection region.
-RCT
What is randomized controlled trial in research?
Randomized controlled trial: (RCT) A study in which people are allocated at random (by
chance alone) to receive one of several clinical interventions. One of these interventions is
the standard of comparison or control. The control may be a standard practice, a placebo
("sugar pill"), or no intervention at all.
Advantages of randomised control trial study design: Disadvantages of randomised control trial
Comparative: study design
o One treatment is directly compared to another to Logistics:
establish superiority. o Power calculation might demand
o This study design can make causal inferences, i.e. it is vast samples size, which require more resources
the strongest empirical evidence of a treatment's efficacy from the investigators
Minimises bias: o Validity requires multiple sites,
o Randomisation minimises allocation bias and selection which will be difficult to manage
bias o Long trial run time may result in the
o Blinding minimises performance bias loss of relevance as practice may have moved
o Double-blinding minimises assessment bias on by the time the trial is published
Minimises confounding factors: selection bias when the study groups are
both known and unknown factors not be widely applicable. Trials which test
o Blocked randomisation makes groups comparable for effectiveness are larger and more expensive
o Statistical test of significance is readily interpretable exclusion criteria; highly controlled setting)
both Type 1 error (where the null hypothesis is incorrectly equipoise: one cannot ethically randomise
rejected) and Type 2 error (where the null hypothesis is patients unless both treatments have equal
o Considered the gold standard: more publishable o Some research cannot be ethically
performed as an RCT (classically, RCT of the
effects of parachutes on the survival of sky-
divers)
- sampling and sampling methods
There are several different sampling techniques available, and they can be subdivided into
two groups: probability sampling and non-probability sampling. In probability (random)
sampling, you start with a complete sampling frame of all eligible individuals from which
you select your sample. In this way, all eligible individuals have a chance of being chosen for
the sample, and you will be more able to generalise the results from your study. Probability
sampling methods tend to be more time-consuming and expensive than non-probability
sampling. In non-probability (non-random) sampling, you do not start with a complete
sampling frame, so some individuals have no chance of being selected. Consequently, you
cannot estimate the effect of sampling error and there is a significant risk of ending up with a
non-representative sample which produces non-generalisable results. However, non-
probability sampling methods tend to be cheaper and more convenient, and they are useful
for exploratory research and hypothesis generation.
Probability Sampling Methods
1. Simple random sampling
In this case each individual is chosen entirely by chance and each member of the population
has an equal chance, or probability, of being selected. One way of obtaining a random sample
is to give each individual in a population a number, and then use a table of random numbers
to decide which individuals to include.1 For example, if you have a sampling frame of 1000
individuals, labelled 0 to 999, use groups of three digits from the random number table to
pick your sample. So, if the first three numbers from the random number table were 094,
select the individual labelled “94”, and so on.
As with all probability sampling methods, simple random sampling allows the sampling error
to be calculated and reduces selection bias. A specific advantage is that it is the most
straightforward method of probability sampling. A disadvantage of simple random sampling
is that you may not select enough individuals with your characteristic of interest, especially if
that characteristic is uncommon. It may also be difficult to define a complete sampling frame
and inconvenient to contact them, especially if different forms of contact are required (email,
phone, post) and your sample units are scattered over a wide geographical area.
2. Systematic sampling
Individuals are selected at regular intervals from the sampling frame. The intervals are chosen
to ensure an adequate sample size. If you need a sample size n from a population of size x,
you should select every x/nth individual for the sample. For example, if you wanted a sample
size of 100 from a population of 1000, select every 1000/100 = 10th member of the sampling
frame.
Systematic sampling is often more convenient than simple random sampling, and it is easy to
administer. However, it may also lead to bias, for example if there are underlying patterns in
the order of the individuals in the sampling frame, such that the sampling technique coincides
with the periodicity of the underlying pattern. As a hypothetical example, if a group of
students were being sampled to gain their opinions on college facilities, but the Student
Record Department’s central list of all students was arranged such that the sex of students
alternated between male and female, choosing an even interval (e.g. every 20th student) would
result in a sample of all males or all females. Whilst in this example the bias is obvious and
should be easily corrected, this may not always be the case.
3. Stratified sampling
In this method, the population is first divided into subgroups (or strata) who all share a
similar characteristic. It is used when we might reasonably expect the measurement of
interest to vary between the different subgroups, and we want to ensure representation from
all the subgroups. For example, in a study of stroke outcomes, we may stratify the population
by sex, to ensure equal representation of men and women. The study sample is then obtained
by taking equal sample sizes from each stratum. In stratified sampling, it may also be
appropriate to choose non-equal sample sizes from each stratum. For example, in a study of
the health outcomes of nursing staff in a county, if there are three hospitals each with
different numbers of nursing staff (hospital A has 500 nurses, hospital B has 1000 and
hospital C has 2000), then it would be appropriate to choose the sample numbers from each
hospital proportionally (e.g. 10 from hospital A, 20 from hospital B and 40 from hospital C).
This ensures a more realistic and accurate estimation of the health outcomes of nurses across
the county, whereas simple random sampling would over-represent nurses from hospitals A
and B. The fact that the sample was stratified should be taken into account at the analysis
stage.
Stratified sampling improves the accuracy and representativeness of the results by reducing
sampling bias. However, it requires knowledge of the appropriate characteristics of the
sampling frame (the details of which are not always available), and it can be difficult to
decide which characteristic(s) to stratify by.
4. Clustered sampling
In a clustered sample, subgroups of the population are used as the sampling unit, rather than
individuals. The population is divided into subgroups, known as clusters, which are randomly
selected to be included in the study. Clusters are usually already defined, for example
individual GP practices or towns could be identified as clusters. In single-stage cluster
sampling, all members of the chosen clusters are then included in the study. In two-stage
cluster sampling, a selection of individuals from each cluster is then randomly selected for
inclusion. Clustering should be taken into account in the analysis. The General Household
survey, which is undertaken annually in England, is a good example of a (one-stage) cluster
sample. All members of the selected households (clusters) are included in the survey.1
Cluster sampling can be more efficient that simple random sampling, especially where a
study takes place over a wide geographical region. For instance, it is easier to contact lots of
individuals in a few GP practices than a few individuals in many different GP practices.
Disadvantages include an increased risk of bias, if the chosen clusters are not representative
of the population, resulting in an increased sampling error.
Non-Probability Sampling Methods
1. Convenience sampling
Convenience sampling is perhaps the easiest method of sampling, because participants are
selected based on availability and willingness to take part. Useful results can be obtained, but
the results are prone to significant bias, because those who volunteer to take part may be
different from those who choose not to (volunteer bias), and the sample may not be
representative of other characteristics, such as age or sex. Note: volunteer bias is a risk of all
non-probability sampling methods.
2. Quota sampling
This method of sampling is often used by market researchers. Interviewers are given a quota
of subjects of a specified type to attempt to recruit. For example, an interviewer might be told
to go out and select 20 adult men, 20 adult women, 10 teenage girls and 10 teenage boys so
that they could interview them about their television viewing. Ideally the quotas chosen
would proportionally represent the characteristics of the underlying population.
Whilst this has the advantage of being relatively straightforward and potentially
representative, the chosen sample may not be representative of other characteristics that
weren’t considered (a consequence of the non-random nature of sampling). 2
3. Judgement (or Purposive) Sampling
Also known as selective, or subjective, sampling, this technique relies on the judgement of
the researcher when choosing who to ask to participate. Researchers may implicitly thus
choose a “representative” sample to suit their needs, or specifically approach individuals with
certain characteristics. This approach is often used by the media when canvassing the public
for opinions and in qualitative research.
Judgement sampling has the advantage of being time-and cost-effective to perform whilst
resulting in a range of responses (particularly useful in qualitative research). However, in
addition to volunteer bias, it is also prone to errors of judgement by the researcher and the
findings, whilst being potentially broad, will not necessarily be representative.
4. Snowball sampling
This method is commonly used in social sciences when investigating hard-to-reach groups.
Existing subjects are asked to nominate further subjects known to them, so the sample
increases in size like a rolling snowball. For example, when carrying out a survey of risk
behaviours amongst intravenous drug users, participants may be asked to nominate other
users to be interviewed.
Snowball sampling can be effective when a sampling frame is difficult to identify. However,
by selecting friends and acquaintances of subjects already investigated, there is a significant
risk of selection bias (choosing a large number of people with similar characteristics or views
to the initial individual identified).
Bias in sampling
There are five important potential sources of bias that should be considered when selecting a
sample, irrespective of the method used. Sampling bias may be introduced when:1
1. Any pre-agreed sampling rules are deviated from
2. People in hard-to-reach groups are omitted
3. Selected individuals are replaced with others, for example if they are difficult to
contact
4. There are low response rates
5. An out-of-date list is used as the sample frame (for example, if it excludes people who
have recently moved to an area)
- What is matching, how is it done, what is demerit of matching too many factors
What is matching in research design?
Matched group design (also known as matched subjects design) is used in experimental
research in order for different experimental conditions to be observed while being able to
control for individual difference by matching similar subjects or groups with each other. This
can best be described using an example.
Advantages of matching
Matching is a useful method to optimize resources in a case control study.
Matching on a factor linked to other factors may automatically control for the confounding
role of those factors (e.g. matching on neighborhood may control for socio-economic
factors).
Matching allows to use a smaller sample size, by preparing the stratified analysis "a priori"
(before the study, at the time of cases and control selection), with smaller sample sizes as
compared to an unmatched sample with stratified analysis made "a posteriori".
Matching avoids a stratified analysis with too many strata, with potentially no case or control,
done to control several confounding factors at the same time. Indeed, in an unmatched case
control study, while we perform logistic regression, or even more simply a stratified analysis,
we might end up with empty strata (no cases or no control in some strata). Matching avoids
this situation.
Disadvantages of matching
The efficiency in data analysis that matching provides is limited by several disadvantages.
The greatest disadvantage of matching is that the effect of matching factor on the occurrence
of the disease of interest cannot be studied anymore. One should therefore limit matching to
factors that are already known to be risk factors for the studied outcome.
If statistical softwares with logistic regression are available, it is possible to control for many
confounding factors during the analysis of the study, and therefore preventing confounding
by matching during the design of the study might not be needed, especially if the study is
including a large population and there are few chances that we will end up with empty strata.
If matching is performed, it must also be taken into account in the statistical analysis, because
a matched OR needs to be calculated, and conditional logistic regression need to be used.
However the study of the matching factor as an effect modifier is still possible if doing a
stratified analysis over several categories of the matching factor. For example when matching
on age, analysis is still feasible within each age stratum created. However to use different age
categories than those used for matching would require a multivariable analysis. Trying to
identify a dose response involving a matching factor would also require a multivariable
model of analysis.
Matching on criteria that are only associated with exposure and not with outcome further
biases the measurement of the effect. In this situation the matching factor is not a
confounding factor and matching would bring the OR towards 1.
Another difficulty occurs when matching on several factors. It then becomes difficult (time
and energy) to logistically identify and recruit controls due the high number of matching
factors (e.g. same age, sex, socio economic status, occupation, etc.). Matching on several
criteria may improve the efficiency of statistical analysis with a reduced sample size but the
difficulties to recruit controls may jeopardize that efficiency. It may also exclude cases for
which no matched controls can be identified. In addition, matching on many criteria increases
the risk of matching on exposure (therefore bringing the OR closer to one). This is sometimes
called overmatching.
One major challenge when matching is to properly define the various strata of the matching
variable. For example when frequency matching on age, we need to make sure that, within
each of the age group created, age is no longer a confounding factor. This is sometimes
called residual confounding. Several analysis with several width of age strata may be tested.
For example, let's suppose we stratify on several age groups 20 years wide (0-19, 20-39, 40-
59, 60-79, 80+). To assess if age is still a confounder within one age group we could further
stratify (by five years age group) and test if age is still a confounding factor inside a 20 years
wide age group. So it may still be important to take account of age as a potential confounder
in a multivariable analysis.
What are the 3 types of t tests?
There are three main types of t-test:
An Independent Samples t-test compares the means for two groups.
A Paired sample t-test compares means from the same group at different times (say,
one year apart).
A One sample t-test tests the mean of a single group against a known mean
RESEARCH METHODOLOGY
Experimental method
Experimental designs,
What are the three principles of experimental design?
Thus, the three principles of experimental design are: replication, to provide an
estimate of experimental error; randomization, to ensure that this estimate is
statistically valid; and. local control, to reduce experimental error by making
the experiment more efficient.
The research design for studying factor of schizophrenia
Prospective
Retrospective
Cohort
What is a cohort study in medical research?
Finding causes
Examples
Limitations
Cohort studies are a type of medical research used to investigate the causes of disease and to
establish links between risk factors and health outcomes.
The word cohort means a group of people. These types of studies look at groups of people.
They can be forward-looking (prospective) or backward-looking (retrospective).
Prospective" studies are planned in advance and carried out over a future period of time.
Retrospective cohort studies look at data that already exist and try to identify risk factors for
particular conditions. Interpretations are limited because the researchers cannot go back and
gather missing data.
These long-term studies are sometimes called longitudinal studies.
Fast facts on cohort studies
Cohort studies typically observe large groups of individuals, recording their exposure
to certain risk factors to find clues as to the possible causes of disease.
They can be prospective studies and gather data going forward, or retrospective
cohort studies, which look at data already collected.
The Nurses' Health Study is one example of a large cohort study, and it has produced
many important links between lifestyle choices and health by following hundreds of
thousands of women across North America.
Such research can also help identify social factors that influence health
Confounding variables
What is a confounding variable in psychology?
Confounding variables are factors other than the independent variable that may
cause a result. In your caffeine study, for example, it is possible that the students who
received caffeine also had more sleep than the control group. Or, the experimental
group may have spent more time overall preparing for the exam
Non-sampling error refers to any deviation between the results of a survey and the truth
which are not caused by the random selecting of observations. That is, non-sampling error is
the total of all forms of error other than sampling error. Common types of non-sampling error
include non-response error, measurement error, interviewer error, adjustment error, and
processing error.
Non-response error
Non-response error refers to errors that are caused by differences between people that
participate in surveys versus people who do not participate in surveys. For example, surveys
that ask people about how they spend their time likely have large amounts of non-response
error, as people who spend their time doing surveys are likely quite different from those who
do not.
Non-response error can be narrowly defined as relating to whether people selected to
participate actually do participate (e.g., the difference between people who opened their email
invitation and completed a survey versus those who did not); or it can be more broadly
defined to include all non-random aspects of sampling (e.g., selection of the list to be used in
the research). Errors relating to whether the lists used in research are representative are also
known as list selection error and coverage error.
Measurement error
Measurement error refers to all the errors relating to the specific measurement of each
sampling unit (as opposed to errors relating to how they were selected to be measured). For
example, these could include confusing question wordings, low-quality data due to
respondent fatigue, and low quality multi-item scales being used to measure abstract
concepts.
Interviewer error
Interviewer error occurs when an interviewer makes an error in how they administer the
survey or record responses. For example, in qualitative research, an interviewer may “lead” a
respondent to a certain answer, and in quantitative research a bored interviewer may choose
to ask a question in words that they regard as superior to those in the questionnaire.
Adjustment error
Adjustment error occurs where the analysis of the data inadvertently adjusts the data in such
a way that it becomes less accurate. The main forms of adjustment error are errors with
weighting, data cleaning, and imputation.
Processing error
Processing error occurs when the processing of the data has caused an error of some kind,
such as when it is incorrectly entered or corrupted.
RESEARCH METHODOLOGY
Experimental method
Experimental designs
What is the definition of experimental research design?
Experimental research is a study that strictly adheres to a scientific research design. It
includes a hypothesis, a variable that can be manipulated by the researcher, and
variables that can be measured, calculated and compared. Most
importantly, experimental research is completed in a controlled environment.
As per Global Burden of Disease report, mental disorders accounts for 13% of total DALYs
lost for Years Lived with Disability (YLD) with depression being the leading cause2 .
Mental Health and Sustainable Development goals Within the health related SDGs, two
targets are directly related to mental health and substance abuse. Target 3.4 “By 2030, reduce
by one third premature mortality from Non communicable diseases through prevention and
treatment and promote mental health and well-being.” Target 3.5 requests that countries:
“Strengthen the prevention and treatment of substance abuse, including narcotic drug abuse
and harmful use of alcohol.” Source: 11
The NMHS was undertaken as a large scale, multi-centred national study on the various
dimensions and characteristics of mental health problems among individuals aged 18 years
and above across 12 Indian states (Figure 1) during 2014 - 16. 2. A pilot study was
undertaken in Kolar district, Karnataka during Jan-Nov 2014, on a sample of 3190
individuals (13 years and above) to examine the feasibility of: conducting the survey, the
proposed sampling methodologies and the use of hand held computing devices for field data
collection. Six well trained data collectors were involved in data collection across 50 clusters
(Villages and urban wards) of the district which provided a crude prevalence of 7.5% for all
mental disorders. The lessons learnt and experience gained helped in developing the NMHS
methodology. 3. The selection of states was based on the availability of an interested and
reliable partner organisation in that state, their willingness to undertake the study and the
availability of screening and diagnostic data collection tools in the vernacular languages
spoken in that state. The states selected were North : Punjab and Uttar Pradesh, South: Tamil
Nadu and Kerala, East: Jharkhand and West Bengal, West: Rajasthan and Gujarat, Central:
Madhya Pradesh and Chhattisgarh and, North-East : Assam and Manip
A multi stage sampling was adopted (District g Taluka g Village / Ward g HH) in each state
and each selected state of India constituted the sampling frame
in total 34802 adults and about 1191 adolescents drawn from 12 states were interviewed