RM&IPR Notes - Unit 1
RM&IPR Notes - Unit 1
Research Problem
Research Problem
It refers to some difficulty which a researcher experiences in the context of either
theoretical or practical situation and wants to obtain a solution for the same.
A research problem can be simply defined as a statement that identifies the problem
or situation to be studied.
Selecting a Problem
Guidelines for selecting a research problem:-
Subject which is overdone should not be chosen
An average researcher must not choose Controversial topics
Too narrow or too vague problems should be avoided
The chosen subject should be familiar and feasible
Significance and Importance of subject must be given attention
Cost and time factor must be kept in mind
Experience, Qualification and Training of the researcher must be according to the problem
in hand
There are few rules that must be kept in mind while defining a research problem. They are-
Technical terms should be clearly defined.
Basic assumptions should be stated.
The criteria for the selection should be provided.
Suitability of the time period and sources of data available must be considered.
The scope of the investigation or the limits must be mentioned.
A very serious impediment facing new researchers is well illustrated by trying to use a
foreign language dictionary to learn what foreign words mean. Until you know "enough" of a
language, you cannot find out what the words mean. Until you know "enough" of a paradigm,
you do not know what unresolved questions remain, or when the paradigm is wrong. "A"
below is generally a person's first exposure to a research project for that reason.
In addition to not knowing what unresolved problems remain, is missing the more
fundamental broader issue underlying any specific behaviour change. When looking at the
world, try to see each functional relationship as only an instance of a more general class of
relationships.
A. Research Problem from Expert
The simplest source of a problem to solve is to have it given to you as a class assignment,
as a directed research project, or as a task while you are an apprentice in someone's lab.
You are told what problem to research and how to do it. This is probably an ideal way to
assure that your first research topic is a good one.
Example: Students in Experimental Psychology were assigned the task of finding out if
social attention made their roommate study more. They were told to measure the amount
of time their roommate studied on days during which they expressed interest in their
roommate's course material as compared to days when they refrained from talking about
academic topics.
B. Research Problem from Folklore
Common beliefs, common sense, or proverbs could be right but on the other hand, they
could also be wrong. You must verify that they are true before considering them as a
source of knowledge. It is possible that some unverified beliefs have the roots of a better
idea and therefore would be a worthy research topic. It is critical to note, however, that
the task of research is not to simply validate or invalidate common sense but rather to
come to understand nature.
Example: It's commonly believed that studying within the two hours preceding a test will
decrease test scores. To research this belief a randomly selected half of a class was told to
study immediately before taking a test while the other half was prohibited from studying
before the test. This research was intended to determine whether or not studying
immediately before a test decreased the points earned.
C. Research Problem from Insight
Sometimes people research an issue simply because it occurred to them and it seemed
important. The systematic development of the idea is lacking. This is "intuitive" or good
guess research. It is risky because you may not be able to get other researchers to
understand why the research is important. It is fun because you get to do what interests
you at the moment. Alternatively, it could be the application of a general rule of thumb or
guessing that a new problem is actually a well-understood function in disguise.
Example: While feeling especially competent after explaining course material to three
friends you realize that orally presenting material may help test performance. You
conducted a study in which material was orally presented before the test on a random
half of the occasions. The research was based on your insightful realization that oral
presentation may increase test performance.
Example: While you were reading about jet lag and its effects on sleep the first night, you
realize that the author failed to control for light cycle. You try stretching either the light
period or stretching the dark period to make up the phase shift. You implement this by
changing the cabin illumination period on various trans-Atlantic flights, and monitoring
the passengers sleep for the next three days.
Area Culture. The culture of the area for which a researcher conducts his research is also
responsible for the selection of research problem.
1. Population Specification
Population specification errors occur when the researcher does not understand who they
should survey. This can be tricky because there are multiple people who might consume the
product, but only one who purchases it, or they may miss a segment looking to purchase in
the future.
Example: Packaged goods manufacturers often conduct surveys of housewives, because
they are easier to contact, and it is assumed they decide what is to be purchased and also do
the actual purchasing. In this situation there often is population specification error. The
husband may purchase a significant share of the packaged goods, and have significant direct
and indirect influence over what is bought. For this reason, excluding husbands from
samples may yield results targeted to the wrong audience.
How to avoid this: Understand who purchases your product and why they buy it. It’s
important to survey the one making the buying decision so you know how to better reach
them.
Here are some example scopes which define a particular problem space. The solution space
appears in parenthesis:
Make sure the drivers in my commercial fleet don’t get us in trouble. (commercial vehicle
speed tracking)
See if I can gain better insights from my engineering/scientific data. (statistical graphical
software)
Figure out why my code isn’t working, to get it working how I’d like. (software API
developers network)
Make sure we have reliable data storage access and future expansion. (data storage
configuration automation)
Decide whether/how I can get a college degree/certificate, given my financial and family
situation. (technical & community college marketing plan)
Guide a business toward success where my employees must drive to customer sites.
(decide whether to build a better dispatch tool)
Decide what to get for lunch. (fast food restaurant)
Data collection is one of the most important stages in conducting a research. You can
have the best research design in the world but if you cannot collect the required data you
will be not be able to complete your project. Data collection is a very demanding job which
needs thorough planning, hard work, patience, perseverance and more to be able to
complete the task successfully. Data collection starts with determining what kind of data
required followed by the selection of a sample from a certain population. After that, you
need to use a certain instrument to collect the data from the selected sample.
TYPES OF DATA
Data are organized into two broad categories: qualitative and quantitative.
Qualitative Data: Qualitative data are mostly non-numerical and usually descriptive or
nominal in nature. This means the data collected are in the form of words and sentences.
Often (not always), such data captures feelings, emotions, or subjective perceptions of
something. Qualitative approaches aim to address the ‘how’ and ‘why’ of a program and
tend to use unstructured methods of data collection to fully explore the topic. Qualitative
questions are open-ended. Qualitative methods include focus groups, group discussions and
interviews. Qualitative approaches are good for further exploring the effects and unintended
consequences of a program. They are, however, expensive and time consuming to
implement. Additionally the findings cannot be generalized to participants outside of the
program and are only indicative of the group involved.
Qualitative data collection methods play an important role in impact evaluation
by providing information useful to understand the processes behind observed results
and assess changes in people’s perceptions of their well-being. Furthermore qualitative
methods can be used to improve the quality of survey-based quantitative evaluations by
helping generate evaluation hypothesis; strengthening the design of survey questionnaires
and expanding or clarifying quantitative evaluation findings. These methods are
characterized by the following attributes -
they tend to be open-ended and have less structured protocols (i.e., researchers may
change the data collection strategy by adding, refining, or dropping techniques or
informants);
they rely more heavily on interactive interviews; respondents may be interviewed
several times to follow up on a particular issue, clarify concepts or check the reliability of
data;
they use triangulation to increase the credibility of their findings (i.e., researchers
rely on multiple data collection methods to check the authenticity of their results);
generally their findings are not generalizable to any specific population, rather each case
study produces a single piece of evidence that can be used to seek general patterns
among different studies of the same issue.
Regardless of the kinds of data involved, data collection in a qualitative study takes a great
deal of time. The researcher needs to record any potentially useful data thoroughly,
accurately, and systematically, using field notes, sketches, audiotapes, photographs and
other suitable means. The data collection methods must observe the ethical principles of
research. The qualitative methods most commonly used in evaluation can be classified in
three broad categories -
In-depth interview
Observation methods
Document review.
Mixed Methods: Mixed methods approach as design, combining both qualitative and
quantitative research data, techniques and methods within a single research
framework. Mixed methods approaches may mean a number of things, i.e. a number of
different types of methods in a study or at different points within a study or using a mixture
of qualitative and quantitative methods. Mixed methods encompass multifaceted
approaches that combine to capitalize on strengths and reduce weaknesses that stem
from using a single research design. Using this approach to gather and evaluate data may
assist to increase the validity and reliability of the research. Some of the common areas
in which mixed-method approaches may be used include –
Initiating, designing, developing and expanding interventions;
Evaluation;
Improving research design; and
Corroborating findings, data triangulation or convergence.
Some of the challenges of using a mixed methods approach include –
Delineating complementary qualitative and quantitative research questions;
Time-intensive data collection and analysis; and
Decisions regarding which research methods to combine.
Mixed methods are useful in highlighting complex research problems such as disparities in
health and can also be transformative in addressing issues for vulnerable or
marginalized populations or research which involves community participation. Using a
mixed-methods approach is one way to develop creative options to traditional or single
design approaches to research and evaluation.
There are many ways of classifying data. A common classification is based upon who
collected the data.
PRIMARY DATA
Data that has been collected from first-hand-experience is known as primary data.
Primary data has not been published yet and is more reliable, authentic and objective.
Primary data has not been changed or altered by human beings; therefore its validity is
greater than secondary data.
Importance of Primary Data: In statistical surveys it is necessary to get information
from primary sources and work on primary data. For example, the statistical records of
female population in a country cannot be based on newspaper, magazine and other
printed sources. A research can be conducted without secondary data but a research based
on only secondary data is least reliable and may have biases because secondary data has
already been manipulated by human beings. One of such sources is old and secondly they
contain limited information as well as they can be misleading and biased.
Sources of Primary Data:
Sources for primary data are limited and at times it becomes difficult to obtain data from
primary source because of either scarcity of population or lack of cooperation.
Following are some of the sources of primary data.
Experiments:
Experiments require an artificial or natural setting in which to perform logical study to
collect data. Experiments are more suitable for medicine, psychological studies, nutrition
and for other scientific studies. In experiments the experimenter has to keep control over
the influence of any extraneous variable on the results.
Survey:
Survey is most commonly used method in social sciences, management, marketing and
psychology to some extent. Surveys can be conducted in different methods.
Questionnaire:
It is the most commonly used method in survey. Questionnaires are a list of questions
either open-ended or close-ended for which the respondents give answers. Questionnaire
can be conducted via telephone, mail, live in a public area, or in an institute, through
electronic mail or through fax and other methods.
Interview:
Interview is a face-to-face conversation with the respondent. In interview the main problem
arises when the respondent deliberately hides information otherwise it is an in depth
source of information. The interviewer can not only record the statements the interviewee
speaks but he can observe the body language, expressions and other reactions to the
questions too. This enables the interviewer to draw conclusions easily.
Observations:
Observation can be done while letting the observing person know that s/he is being
observed or without letting him know. Observations can also be made in natural settings as
well as in artificially created environment.
Advantages of Using Primary Data
The investigator collects data specific to the problem under study.
There is no doubt about the quality of the data collected (for the investigator).
If required, it may be possible to obtain additional data during the study period.
Disadvantages of Using Primary Data
1. The investigator has to contend with all the hassles of data collection-
deciding why, what, how, when to collect;
getting the data collected (personally or through others);
getting funding and dealing with funding agencies;
ethical considerations (consent, permissions, etc.).
2. Ensuring the data collected is of a high standard-
all desired data is obtained accurately, and in the format it is required in;
there is no fake/ cooked up data;
unnecessary/ useless data has not been included.
3. Cost of obtaining the data is often the major expense in studies.
SECONDARY DATA
Data collected from a source that has already been published in any form is
called as secondary data. The review of literature in any research is based on secondary
data. It is collected by someone else for some other purpose (but being utilized by the
investigator for another purpose). For examples, Census data being used to analyze the
impact of education on career choice and earning.
Common sources of secondary data for social science include censuses,
organizational records and data collected through qualitative methodologies or qualitative
research. Secondary data is essential, since it is impossible to conduct a new survey that
can adequately capture past change and/or developments.
Sources of Secondary Data:
The following are some ways of collecting secondary data –
Books
Records
Biographies
Newspapers
Published censuses or other statistical data
Data archives
Internet articles
Research articles by other researchers (journals)
Databases, etc.
Importance of Secondary Data:
Secondary data can be less valid but its importance is still there. Sometimes it is difficult to
obtain primary data; in these cases getting information from secondary sources is easier and
possible. Sometimes primary data does not exist in such situation one has to confine the
research on secondary data. Sometimes primary data is present but the respondents are not
willing to reveal it in such case too secondary data can suffice. For example, if the research
on the psychology of transsexuals first it is difficult to find out transsexuals and second they
may not be willing to give information you want for your research, so you can collect data
from books or other published sources. A clear benefit of using secondary data is that much
of the background work needed has already been carried out. For example, literature
reviews, case studies might have been carried out, published texts and statistics could
have been already used elsewhere, media promotion and personal contacts have also been
utilized. This wealth of background work means that secondary data generally have a pre-
established degree of validity and reliability which need not be re-examined by the
researcher who is re-using such data. Furthermore, secondary data can also be helpful in the
research design of subsequent primary research and can provide a baseline with which the
collected primary data results can be compared to. Therefore, it is always wise to begin
any research activity with a review of the secondary data.
Advantages of Using Secondary Data
No hassles of data collection.
It is less expensive.
The investigator is not personally responsible for the quality of data (‘I didn’t do it’).
Disadvantages of Using Secondary Data
The data collected by the third party may not be a reliable party so the reliability and
accuracy of data go down.
Data collected in one location may not be suitable for the other one due variable
environmental factor.
With the passage of time the data becomes obsolete and very old.
Secondary data collected can distort the results of the research. For using secondary
data a special care is required to amend or modify for use.
Secondary data can also raise issues of authenticity and copyright.
Keeping in view the advantages and disadvantages of sources of data requirement of the
research study and time factor, both sources of data i.e. primary and secondary data have
been selected. These are used in combination to give proper coverage to the topic.
PROCESS ANALYSIS
A step-by-step breakdown of the phases of a process, used to convey the inputs, outputs, and
operations that take place during each phase. A process analysis can be used to improve
understanding of how the process operates, and to determine potential targets for process
improvement through removing waste and increasing efficiency. Inputs may be materials,
labor, energy, and capital equipment. Outputs may be a physical product (possibly used as
an input to another process) or a service. Processes can have a significant impact on the
performance of a business, and process improvement can improve a firm’s competitiveness.
The first step to improving a process is to analyze it in order to understand the activities,
their relationships, and the values of relevant metrics. Process analysis generally involves
the following tasks-
Define the process boundaries that mark the entry points of the process inputs and the
exit points of the process outputs.
Construct a process flow diagram that illustrates the various process activities and their
interrelationships.
Determine the capacity of each step in the process. Calculate other measures of interest.
Identify the bottleneck, that is, the step having the lowest capacity.
Evaluate further limitations in order to quantify the impact of the bottleneck.
Use the analysis to make operating decisions and to improve the process.
Process Analysis Tools
When you want to understand a work process or some part of a process, these tools can help
Flowchart: A picture of the separate steps of a process in sequential order, including
materials or services entering or leaving the process (inputs and outputs), decisions that
must be made, people who become involved, time involved at each step and/or process
measurements.
Failure Mode Effects Analysis (FMEA): A step-by-step approach for identifying all possible
failures in a design, a manufacturing or assembly process, or a product or service;
studying the consequences, or effects, of those failures; and eliminating or reducing
failures, starting with the highest-priority ones.
Mistake-proofing: The use of any automatic device or method that either makes it
impossible for an error to occur or makes the error immediately obvious once it has
occurred.
Spaghetti Diagram: A spaghetti diagram is a visual representation using a continuous flow
line tracing the path of an item or activity through a process. The continuous flow line
enables process teams to identify redundancies in the work flow and opportunities to
expedite process flow.
Process Flow Diagram The process boundaries are defined by the entry and exit points of
inputs and outputs of the process. Once the boundaries are defined, the process flow
diagram (or process flowchart ) is a valuable tool for understanding the process using
graphic elements to represent tasks, flows, and storage. The following is a flow diagram for a
simple process having three sequential activities-
Link analysis is a data analysis technique used in network theory that is used to evaluate the
relationships or connections between network nodes. These relationships can be between
various types of objects (nodes), including people, organizations and even transactions. Link
analysis is essentially a kind of knowledge discovery that can be used to visualize data to
allow for better analysis, especially in the context of links, whether Web links or relationship
links between people or between different entities. Link analysis has been used for
investigation of criminal activity (fraud detection, counterterrorism, and intelligence),
computer security analysis, search engine optimization, market research and medical
research.
Link analysis is literally about analyzing the links between objects, whether they are
physical, digital or relational. This requires diligent data gathering. For example, in the case
of a website where all of the links and backlinks that are present must be analyzed, a tool has
to sift through all of the HTML codes and various scripts in the page and then follow all the
links it finds in order to determine what sort of links are present and whether they are active
or dead. This information can be very important for search engine optimization, as it allows
the analyst to determine whether the search engine is actually able to find and index the
website. In networking, link analysis may involve determining the integrity of the connection
between each network node by analyzing the data that passes through the physical or virtual
links. With the data, analysts can find bottlenecks and possible fault areas and are able to
patch them up more quickly or even help with network optimization.
Link analysis has three primary purposes –
Find matches for known patterns of interests between linked objects.
Find anomalies by detecting violated known patterns.
Find new patterns of interest (for example, in social networking and marketing and
business intelligence).
Instrument, Validity, Reliability
Instrument is the general term that researchers use for a measurement device (survey, test,
questionnaire, etc.). To help distinguish between instrument and instrumentation, consider
that the instrument is the device and instrumentation is the course of action (the process of
developing, testing, and using the device).
Instruments fall into two broad categories, researcher-completed and subject-completed,
distinguished by those instruments that researchers administer versus those that are
completed by participants. Researchers chose which type of instrument, or instruments, to
use based on the research question. Examples are listed below:
Researcher-completed Instruments Subject-completed Instruments
Rating scales Questionnaires
Interview schedules/guides Self-checklists
Tally sheets Attitude scales
Flowcharts Personality inventories
Performance checklists Achievement/aptitude tests
Time-and-motion logs Projective devices
Observation forms Sociometric devices
Usability refers to the ease with which an instrument can be administered, interpreted by
the participant, and scored/interpreted by the researcher. Example usability problems
include:
1. Students are asked to rate a lesson immediately after class, but there are only a few
minutes before the next class begins (problem with administration).
2. Students are asked to keep self-checklists of their after school activities, but the
directions are complicated and the item descriptions confusing (problem with
interpretation).
3. Teachers are asked about their attitudes regarding school policy, but some questions are
worded poorly which results in low completion rates (problem with
scoring/interpretation).
Validity and reliability concerns (discussed below) will help alleviate usability issues. For
now, we can identify five usability considerations:
1. How long will it take to administer?
2. Are the directions clear?
3. How easy is it to score?
4. Do equivalent forms exist?
5. Have any problems been reported by others who used it?
It is best to use an existing instrument, one that has been developed and tested numerous
times, such as can be found in the Mental Measurements Yearbook. We will turn to why
next.
Validity is the extent to which an instrument measures what it is supposed to measure and
performs as it is designed to perform. It is rare, if nearly impossible, that an instrument be
100% valid, so validity is generally measured in degrees. As a process, validation involves
collecting and analyzing data to assess the accuracy of an instrument. There are numerous
statistical tests and measures to assess the validity of quantitative instruments, which
generally involves pilot testing. The remainder of this discussion focuses on external validity
and content validity.
External validity is the extent to which the results of a study can be generalized from a
sample to a population. Establishing eternal validity for an instrument, then, follows directly
from sampling. Recall that a sample should be an accurate representation of a population,
because the total population may not be available. An instrument that is externally valid
helps obtain population generalizability, or the degree to which a sample represents the
population.
Content validity refers to the appropriateness of the content of an instrument. In other
words, do the measures (questions, observation logs, etc.) accurately assess what you want
to know? This is particularly important with achievement tests. Consider that a test
developer wants to maximize the validity of a unit test for 7th grade mathematics. This
would involve taking representative questions from each of the sections of the unit and
evaluating them against the desired outcomes.
Reliability can be thought of as consistency. Does the instrument consistently measure what
it is intended to measure? It is not possible to calculate reliability; however, there are four
general estimators that you may encounter in reading research:
1. Inter-Rater/Observer Reliability: The degree to which different raters/observers give
consistent answers or estimates.
2. Test-Retest Reliability: The consistency of a measure evaluated over time.
3. Parallel-Forms Reliability: The reliability of two tests constructed the same way, from
the same content.
4. Internal Consistency Reliability: The consistency of results across items, often measured
with Cronbach’s Alpha.