Bed 212 Course Material
Bed 212 Course Material
SCHOOL OF EDUCATION
COURSE AIMS
This course aims at producing competent business educators who will be versed in
organizational and research data processing in order to foster time and effective information that
will be used in decision making. In order to enable you meet the above aims, modules
constituting of units have been produced for your study. Apart from meeting the aims of the
course as a whole, each course unit consists of learning objectives which are intended to ensure
your learning effectiveness.
COURSE OBJECTIVES
The course objectives are meant to enable you achieve/acquire the following:
1) Gain in-depth knowledge of data processing and its functions in business organisations
and educational settings.
2) Appreciate the historical perspective of data processing and how it evolves from manual
to electronically based data processing.
3) Acquire skills in generating researchable problems and derive relevant research questions
and hypotheses.
4) Understanding of the basic research designs to employ when conducting business
education studies.
5) Appreciate various research data collection tools that are available to various business
and educational projects
6) Demonstrate skills in utilizing appropriate statistical tools in analyzing research data and
giving relevant interpretations to results generated.
7) Write comprehensive research reports after acquiring knowledge of the components of
report writing.
COURSE MATERIALS
The major components of this course are
1. Course Guide
2. Study units
3. Text books
4. CDS
5. Practice In-text questions
6. Tutor Assignment file
7. E- Facilitation
STUDY UNITS
The breakdown of the three modules and study units are as follows: Every unit contains a list of
references and further reading. Try as much as you can to get the listed books and further reading
materials. You are also expected to approach the internet for further related reading materials.
This is to widen as well as deepen the depth of understanding of this course
ASSESSMENT
Your assessment will be based on tutor-marked assignments (TMAs) and a final examination
which you will write at the end of the course. You are required to complete four TMAs which
will contain 20 multiple choice test items each.
1.0 INTRODUCTION
Data Processing (data processing) is a data collection, storage, retrieval, processing,
transformation and transmission. Data is facts, concepts, or instructions of a form of expression,
can be manual or automated device for processing in form of digital data, text, graphics or
sounds. Data are explained and, after giving a certain sense, they become information. The basic
purpose of data processing from a large number, possibly chaotic, difficult to understand and
extract the data derived for some specific people who are valuable and meaningful data. Data
processing is a systems engineering and automatic control of basic aspects. Data processing is
applied throughout the social production and social life in various fields. Data processing
technology and its application breadth and depth greatly affect the process of development of
human society. No one be effective in any life’s endeavour without the support of data
processing software, data processing software includes: writing process for various programming
languages and compilers, data management, file systems and database systems, and various
methods of data processing software package. To ensure data security and reliable, a range of
data security technologies and management is very essential.
2.0 OBJECTIVES
At the end of this unit, you should be able to
You must have performed many daily operations that need your skills in data processing. As
you wake up in the morning, you think about number of activities to be scheduled for the day,
the amount of money at your disposal to execute the activities and number of people to meet to
make you achieve your goals for the day. Relevant data are manipulated to get things done.
3.1.1 Data
Data are any facts, numbers, or text that can be processed by a computer. Today, organizations
are accumulating vast and growing amounts of data in different formats and different databases.
This includes:
• operational or transactional data such as, sales, cost, inventory, payroll, and accounting
• non-operational data, such as industry sales, forecast data, and macro economic data
• meta data - data about the data itself, such as logical database design or data dictionary
definitions
3.1.2 Information
The patterns, associations, or relationships among all this data can provide information. For
example, analysis of retail point of sale transaction data can yield information on which products
are selling and when. The need for converting facts into useful information is not a phenomenon
of modern life. Throughout history, and even prehistory, people have found it necessary to sort
data into forms that were easier to understand. For example, the ancient Egyptians recorded the
ebb and flow of the Nile River and used this information to predict yearly crop yields. Today
computers convert data about land and water into recommendations to farmers on crop planting.
Mechanical aids to computation were developed and improved upon in Europe, Asia, and
America throughout the seventeenth, eighteenth, and nineteenth centuries. Modern computers are
marvels of an electronics technology that continues to produce smaller, cheaper, and more pow-
erful components.
3.1.3 Data Processing
Data are a collection of facts — unorganized but able to be organized into useful information.
Processing is a series of actions or operations that convert inputs into outputs. When we speak of
data processing, the input is data, and the output is useful information. What then is data
processing? So, we can define data processing as a series of actions or operations that converts
data into useful information.
According to Carl (1996) data processing is, broadly, "the collection and manipulation of items
of data to produce meaningful information. In this sense it can be considered a subset of
information processing, "the change (processing) of information in any manner detectable by an
observer. The term is often used more specifically in the context of a business or other
organization to refer to the class of commercial data processing applications.
Data processing (numerical and non-numerical) includes the analysis of various, sorting,
calculating, editing, processing and handling data. The increasing popularity of computers in the
field of computer applications, a small proportion of numerical calculation by computer data
processing for information management has become a major application. Such as the side of the
draw chart management, warehouse management, accounting management, transportation
management, IT management, office automation. Geographical data in a large number of
existing data in the natural environment (land, water, climate, biological and other resource
data), there are a large number of socio-economic data (population, transport, industry and
agriculture, etc.), often require comprehensive data processing. Therefore, the need to establish
geographic database, the system to collate and store geographic data to reduce redundancy, the
development of data processing software, full use of database technology for data management
and processing. Research data also are processed to answer research questions and test
hypotheses.
For data processing business website: As the site visits is very large, making a number of
professional data analysis, data cleansing often have targeted, that is not related to data,
important data such as disposed of. Then, the data related to sub-classification of the segments,
you can choose, depending on the analysis of demand pattern analysis techniques, such as path
analysis, interest in association rules, clustering and so on.
We use the term data processing system to include the resources that are used to accomplish the
processing of data. There are four types of resources: people, materials, facilities, and equipment.
People provide input to computers, operate them, and use their output. Materials, such as boxes
of paper and printer ribbons, are consumed in great quantity. Facilities are required to house the
computer equipment, people and materials.
Ordinal - (categories have inherent order). E.g. Job grade, age groups
Quantitative (Interval/Ratio)
Continuous - Can, in theory, take any value in a range, although necessarily recorded to a
IN‐TEXT QUESTIONS:
Five basic operations are characteristic of all data processing systems: inputting, storing,
processing, outputting, and controlling. They are defined as follows.
Inputting is the process of entering data, which are collected facts, into a data processing system.
Storing is saving data or information so that they are available for initial or for additional
processing. Processing represents performing arithmetic or logical operations on data in order to
convert them into useful information. Outputting is the process of producing useful information,
such as a printed report or visual display. Controlling is directing the manner and sequence in
which all of the above operations are performed.
It is known that data, once entered, are organized and stored in successively more comprehensive
groupings. Generally, these groupings are called a data storage hierarchy. The general groupings
of any data storage hierarchy are as follows:
1) Characters, which are all written language symbols: letters, numbers, and special symbols. 2)
Data elements, which are meaningful collections of related characters. Data elements are also
called data items or fields. 3) Records, which are collections of related data elements. 4) Files,
which are collections of related records. A set of related files is called a data base or a data bank.
You also need to familiarize yourself with these data processing terms that you will come across
often during study of this course.
Now that we have discussed functions and components of data processing, let us examine its
advantages. Computer-oriented data processing systems or just computer data processing
systems are not designed to imitate manual systems. They should combine the capabilities of
both humans and computers. Computer data processing systems can be designed to take
advantage of four capabilities of computers.
Accuracy. Once data have been entered correctly into the computer component of a data
processing system, the need for further manipulation by humans is eliminated, and the possibility
of error is reduced. Computers, when properly programmed, are also unlikely to make
computational errors. Of course, computer systems remain vulnerable to the entry by humans of
invalid data.
Capacity of storage. Computers are able to store vast amounts of information, to organize it,
and to retrieve it in ways that are far beyond the capabilities of humans. The amount of data that
can be stored on devices such as magnetic discs is constantly increasing. All the while, the cost
per character of data stored is decreasing.
Speed. The speed, at which computer data processing systems can respond, adds to their
value. For example, the travel reservations system mentioned above would not be useful if cli-
ents had to wait more than a few seconds for a response. The response required might be a
fraction of a second. Thus, an important objective in the design of computer data processing
systems is to allow computers to do what they do best and to free humans from routine, error-
prone tasks. The most cost-effective computer data processing system is the one that does the job
effectively and at the least cost. By using computers in a cost-effective manner, we will be better
able to respond to the challenges and opportunities of our post-industrial, information-dependent
society.
processing;
10. Storing j) the process of entering collected into a data processing system;
4.0 Conclusion
The unit 1 you have studied had exposed you to meaning of data, information and data
processing.
5.0 Summary
Data are any facts, numbers, or text that can be processed by a computer. Data
processing (numerical and non-numerical) includes the analysis of various, sorting, calculating,
editing, processing and handling data. Advantages of data processing involve accuracy, easy of
communication, capacity of storage and speed.
Carl, F. (1996). Data Processing and Information Technology (10th ed.). London: Thomson.
Starck, J.L., Murtagh, F. & Bijaoui, A (1998). Image Processing and Data Analysis: The
Multiscale Approach. Washington. Amazon
UNIT 2: HISTORICAL PERSPECTIVES AND APPLICATION
OF DATA PROCESSING
CONTENTS
1.0 Introduction
2.0 Objectives
4.0 Conclusion
5.0 Summary
1.0 INTRODUCTION
In unit 1, you came across some of the key terms that you need to understand to
continue in this course. You have studied that data processing involves inputting,
processing and outputting. We shall in this unit intimate ourselves with the historical
perspectives of data processing and necessary steps to effective data processing would
also be discussed.
2.0 OBJECTIVES
3.0 CONTENTS
The 1850 United States Census schedule was the first to gather data by individual rather than
household. A number of questions could be answered by making a check in the appropriate box
on the form. From 1850 through 1880 the Census Bureau employed "a system of tallying, which,
by reason of the increasing number of combinations of classifications required, became
increasingly complex. Only a limited number of combinations could be recorded in one tally, so
it was necessary to handle the schedules 5 or 6 times, for as many independent tallies."[4] "It took
over 7 years to publish the results of the 1880 census"[5] using manual processing methods.
3.2 Applications
3.2.1 Commercial data processing
Commercial data processing involves a large volume of input data, relatively few computational
operations, and a large volume of output. For example, an insurance company needs to keep
records on tens or hundreds of thousands of policies, print and mail bills, and receive and post
payments.
Data is an integral part of all business processes. It is the invisible backbone that supports all the
operations and activities within a business. Without access to relevant data, businesses would get
completely paralyzed. This is because quality data helps formulate effective business strategies
and fruitful business decisions. Therefore, the quality of data should be maintained in good
condition in order to facilitate smooth business proceedings. In order to enhance business
proceedings, data should be made available in all possible forms in order to increase the
accessibility of the same.
Data processing refers to the process of converting data from one format to another. It transforms
plain data into valuable information and information into data. Clients can supply data in a
variety of forms, be it .xls sheets, audio devices, or plain printed material. Data processing
services take the raw data and process it accordingly to produce sensible information. The
various applications of data processing can convert raw data into useful information that can be
used further for business processes.
Companies and organizations across the world make use of data processing services in order to
facilitate their market research interests. Data consists of facts and figures, based on which
important conclusions can be drawn. When companies and organizations have access to useful
information, they can utilize it for strategizing powerful business moves that would eventually
increase the company revenue and decrease the costs, thus expanding the profit margins. Data
processing ensures that the data is presented in a clean and systematic manner and is easy to
understand and be used for further purposes.
Editing
There is a big difference between data and useful data. While there are huge volumes of data
available on the internet, useful data has to be extracted from the huge volumes of the same.
Extracting relevant data is one of the core procedures of data processing. When data has been
accumulated from various sources, it is edited in order to discard the inappropriate data and
retain relevant data.
Coding
Even after the editing process, the available data is not in any specific order. To make it more
sensible and usable for further use, it needs to be aligned into a particular system. The method of
coding ensures just that and arranges data in a comprehendible format. The process is also
known as netting or bucketing.
Data Entry
After the data has been properly arranged and coded, it is entered into the software that performs
the eventual cross tabulation. Data entry professionals do the task efficiently.
IN-TEXT QUESTIONS
3. Data processing ensures that the data is presented in a clean and systematic
manner and is easy to understand and be used for further purposes. True or
False
Validation
After the cleansing phase, comes the validation process. Data validation refers to the process of
thoroughly checking the collected data to ensure optimal quality levels. All the accumulated data
is double checked in order to ensure that it contains no inconsistencies and is utterly relevant.
Tabulation
This is the final step in data processing. The final product i.e. the data is tabulated and arranged
in a systematic format so that it can be further analyzed.
All these processes make up the complete data processing activity which ensures the said data is
available for access.
4.0 CONCLUSION
Data processing had passed through stages of development in which electronic data
processing had taken over all forms of manual data processing. Data processing
include the steps of data entry, editing, coding, validation and tabulation.
5.0 SUMMARY
In this unit you have learnt about the historical evolution of data processing and the
steps involved in data processing. The discussion in this unit had also helped you to
appreciate how data processing can be carried out in various organisation.
Evolvement of data processing comprised manual, automated and electronic
operations to put data into meaning form. Data processing is basically applied to data
analysis and for commercial purposes. You will come across hard ware components
of data processing in unit 3.
6.0 TUTOR-MARKED ASSIGNMENT
ii. Describe 5 steps that are essential for effective data processing
1.0 Introduction
2.0 Objectives
4.0 Conclusion
5.0 Summary
1.0 INTRODUCTION
Now that you have covered the concept of data processing and its historical perspectives
in units 1 and 2, it is necessary to discuss the hardware supports in data processing. In this unit
we shall look at input, processing and output devices for effective data processing in various
organisations.
2.0 OBJECTIVES
Whenever a computer is used it must work its way through three basic stages before any task
can be completed. These are input, processing and output. A Computer works through these
stages by running a program. A program is a set of step-by-step instructions which tells the
computer exactly what to do with the input in order to produce the required output.
3.1.1 Input
The input stage of computing is concerned with getting the data needed by the program into the
computer. Input devices are used to do this. The most commonly used input devices are the
mouse and the keyboard.
3.1.2 Processing
The program contains instructions about what to do with the input. During the processing stage
the compute follows these instructions using the data which has just been input. What the
computer produces at the end of this stage, the output, will only be as good as the instructions
given in the program. In other words if garbage has been put in to the program, garbage is what
will come out of the computer. This is known as GIGO, or Garbage In Garbage Out.
3.1.3 Output
The output stage of computing is concerned with giving out processed data as information in a
form that is useful to the user. Output devices are used to do this. The most commonly used
output devices are the screen, which is also called a monitor or VDU and the printer.
This is the 'brain' of the computer. It is where all the searching, sorting, calculating and decision
making takes place. The CPU collects all of the raw data from various input devices (such a
keyboard or mouse) and converts it into useful information by carrying out software instructions.
The result of all that work is then sent to output devices such as monitors and printers.
The CPU is a microprocessor - a silicon chip - composed of tiny electrical switches called
'transistors'. The speed at which the processor carries out its operations is measured in megahertz
(MHz) or Gigahertz (GHz). The higher the number of MHz the faster the computer can process
information. A common CPU today runs at around 3 GHz or more.
The Intel Pentium processor and the Athlon are examples of a CPU.
INPUT DEVICES
OUTPUT
DEVICES
The Control Unit (CU) co-ordinates the work of the whole computer system. It has three main
jobs:
1. It controls the hardware attached to the system. The Control Unit monitors the hardware to
make sure that the commands given to it by the current program are activated.
2. It controls the input and output of data, so all the signals go to the right place at the right time.
The Immediate Access Store (IAS) holds the data and programs needed at that instant by the
Control Unit. The CPU reads data and programs kept on the backing storage and store them
temporarily in the IAS's memory. The CPU needs to do this because Backing Store is much too
slow to be able to run data and programs from directly. For example, lets pretend that a modern
CPU was slowed down to carry out one instruction in 1 second, then the hard disk (i.e. Backing
Store) would take 3 months to supply the data it needs!
So the trick is to call in enough of the data and programs into fast Immediate Access Store
memory so as to keep the CPU busy.
1. Arithmetic part - does exactly what you think it should - it does the calculations on data such
as 3 + 2.
2. Logic part - This section deals with carrying out logic and comparison operations on data. For
example working out if one data value is bigger than another data value.
Due to a constant research in the computer hardware we have a large number of input devices
recall that before data can be processed by the computer they must be translated into machine
readable form and entered into the computer by an input device. Here we will introduce a variety
of input devices.
3.2.4.1 Keyboard
The keyboard is the most widely used input device and is used to enter data or commands to the
computer. It has a set of alphabet keys, a set of digit keys, and various function keys and is
divided into four main areas:
• Cursor movement and editing keys between the main section and the numeric keypad.
The layout of the letters on a keyboard is standard across many countries and is called a
QWERTY keyboard. The name comes from the first six keys on the top row of the alphabetic
characters. Some keyboards come with added keys for using the Internet and others have an
integrated wrist support. Ergonomic keyboards have been developed to reduce the risk of
repetitive strain injury to workers who use keyboards for long periods of time.
The computer's processor scans the keyboard hundreds of times per second to see if a key has
been pressed. When a key is pressed, a digital code is sent to the Central Processing Unit (CPU).
This digital code is translated into ASCII code (American Standard Code of Information
Interchange).
For example, pressing the 'A' key produces the binary code 01100001 representing the lower
case letter 'a'. Holding down the shift key at the same time produces the binary code 01000001
representing the upper case letter 'A'.
Advantages:
• Most computers have this device attached to it
Disadvantages:
• It can be very time consuming to enter data using a keyboard, especially if you are not a skilled
typist.
• It is very difficult to enter some data, for example, details of diagrams and pictures.
• It is very slow to access menus and not flexible when you want to move objects around the
screen
• Difficult for people unable to use keyboards through paralysis or muscular disorder.
3.2.4.2 Mouse
A mouse is the most common pointing device that you will come across. It enables you to
control the movement and position of the on-screen cursor by moving it around on the desk.
Buttons on the mouse let you select options from menus and drag objects around the screen.
Pressing a mouse button produces a 'mouse click'. You might have heard the expressions 'double
click', 'click and drag' and 'drag and drop'.
Most mice use a small ball located underneath them to calculate the direction that you are
moving the mouse in. The movement of the ball causes two rollers to rotate inside the mouse;
one records the movement in a north-south direction and the other records the east-west
movement. The mouse monitors how far the ball turns and in what direction and sends this
information to the computer to move the pointer.
Advantages:
Disadvantages:
• They need a flat space close to the computer.
• The mouse cannot easily be used with laptop, notebook or palmtop computers. (These need a
tracker ball or a touch sensitive pad called a touch pad).
3.2.4.3 Trackball
A tracker ball is like an upside down mouse with the ball on top. Turning the ball with your hand
moves the pointer on the screen. It has buttons like a standard mouse, but requires very little
space to operate and is often used in conjunction with computer aided design. You will often find
a small tracker ball built into laptop computers in place of the conventional mouse.
Advantages:
• Ideal for use where flat space close to the computer is limited.
• Can be useful with laptops as they can be built into the computer keyboard or clipped on.
Disadvantages:
• Not supplied as standard so an additional cost and users have to learn how to use them
3.2.4.4 Joystick
A Joystick is similar to a tracker ball in operation except you have a stick which is moved rather
than a rolling ball. Joysticks are used to play computer games. You can
move a standard joystick in any one of eight directions. The joystick tells the computer in which
direction it is being pulled and the computer uses this information to
(for example) move a r acing car on screen. A joystick may also have several buttons which can
be pressed to trigger actions such as firing a missile.
Advantages:
Disadvantages:
• Some people find the joystick difficult to control rather than other point and click devices. This
is probably because more arm and wrist movement is required to control the pointer than with a
mouse or tracker ball.
• Joysticks are not particularly strong and can break easily when used with games software.
criss-cross the computer screen. When you touch the screen with your finger, the rays are
blocked and the computer 'senses' where you have pressed. Touch screens can be used to choose
options which are displayed on the screen.
Touch screens are easy to use and are often found as input devices in public places such as
museums, building societies (ATMs), airports or travel agents. However, they are not commonly
used elsewhere since they are not very accurate, tiring to use for a long period and are more
expensive than alternatives such as a mouse.
Advantages:
• Easy to use
• Software can alter the screen while it is running, making it more flexible that a concept
keyboard with a permanent overlay
• No extra peripherals are needed apart from the touch screen monitor itself.
• No experience or competence with computer systems are needed to be able to use it.
Disadvantages:
• Not very accurate, selecting detailed objects can be difficult with fingers
• Touch screens are not robust and can soon become faulty.
A digital camera looks very similar to a traditional camera. However, unlike photographic
cameras, digital cameras do not use film. Inside a digital camera is an array of light sensors.
When a picture is taken, the different colors that make up the picture are converted into digital
signals (binary) by sensors placed behind the lens.
Most digital cameras let you view the image as soon as you have taken the picture and, if you
don't like what you see, it can be deleted. The image can then be stored in the camera's RAM or
on a floppy disk. Later, the pictures can be transferred onto a computer for editing using photo
imaging software. The amount of memory taken up by each picture depends on its resolution.
The resolution is determined by the number of dots which make up the picture: the greater the
number of dots which make up the picture, the clearer the image. However, higher resolution
pictures take up more memory (and are more expensive).
Resolution range from about 3 million (or Mega) pixels up to 12 Mega pixels Digital cameras are
extremely useful for tasks such as producing newsletters. There is often a digital camera built
into mobile phones that operates in exactly the same way as a standard one.
Advantages:
• Images can be incorporated easily into documents, sent by e-mail or added to a website.
Disadvantages:
• Images often have to be compressed to avoid using up too much expensive memory
• When they are full, the images must be downloaded to a computer or deleted before any more
can be taken.
3.2.4.7 Scanner
A scanner is another way in which we can capture still images or text to be stored and used on a
computer. Images are stored as 'pixels'.
A scanner works by shining a beam of light on to the surface of the object you are scanning. This
light is reflected back on to a sensor that detects the color of the light. The reflected light is then
digitized to build up a digital image. Scanner software usually allows you to choose between a
high resolution (very high quality images taking up a lot of memory) and lower resolutions.
Special software can also be used to convert images of text into actual text data which can be
edited by a word processor. This software is called an "Optical Character Reader" or OCR.
• Flatbed Scanner
• Handheld Scanner
The most popular type of scanner is the flatbed. It works in a similar way to a photocopier.
Flatbed scanners can scan larger images and are more accurate than handheld scanners.
Handheld scanners are usually only a few inches wide and are rolled across the document to be
scanned. They perform the same job but the amount of information that can be scanned is limited
by the width of the scanner and the images produced are not of the same quality as those
produced by flatbed scanners.
Advantages:
• Flat-bed scanners are very accurate and can produce images with a far higher resolution than a
digital camera
• Any image can be converted from paper into digital format and later enhanced and used in
other computer documents.
Disadvantages:
• The quality of the final image depends greatly upon the quality of the original document.
Graphics tablets are often used by graphics designers and illustrators. Using a graphics tablet a
designer can produce much more accurate drawings on the screen than they could with a mouse
or other pointing device. A graphics tablet consists of a flat pad (the tablet) on which you draw
with a special pen. As you draw on the pad the image is created on the screen. By using a
graphics tablet a designer can produce very accurate on-screen drawings.
Drawings created using a graphics tablet can be accurate to within hundredths of an inch. The
'stylus' or pen that you use may have buttons on it that act like a set of mouse buttons.
Sometimes, instead of a stylus a highly accurate mouse-like device called a puck is used to draw
on the tablet.
Advantages:
• In a design environment where it is more natural to draw diagrams with pencil and paper, it is
an effective method of inputting the data into the computer.
Disadvantages:
However, output can also be to a modem, a plotter, speakers, a computer disk, another computer
or even a robot.
3.2.5.1 Monitor
A Monitor (or "screen") is the most common form of output from a computer. It displays
information in a similar way to that shown on a television screen. On a typical computer the
monitor may measure 17 inches (43 cm) across its display area. Larger monitors make working
at a computer easier on the eyes. Of course the larger the screen, the higher its cost! Typical
larger sizes are 19 inch, 20 inch and 21 inches.
Part of the quality of the output on a monitor depends on what resolution it is capable of
displaying. Other factors include how much contrast it has, its viewing angle and how fast does it
refresh the screen. For example a good computer game needs a fast screen refresh so you can see
all the action. The picture on a monitor is made up of thousands of tiny colored dots called
pixels. The quality and detail of the picture on a monitor depends on the number of pixels that it
can display. The more dense the pixels the greater the clarity of the screen image.
A PC monitor contains a matrix of dots of Red, Green and Blue known as RGB. These can be
blended to display millions of colors.
R + B = M (magenta)
B + G = C (cyan)
G + R = Y (yellow)
R + G + B = W (white)
The two most common types of monitor are a cathode-ray tube (CRT) monitor and a liquid
crystal display (LCD).
This is smaller and lighter than the CRT (see below), which makes them ideal for use with
portable laptops, PDAs and Palmtops. Even desktop computers are using them now that their
price has become comparable to CRT monitors.
Liquid Crystal is the material used to create each pixel on the screen. The material has a special
property - it can 'polarize' light depending on the electrical charge across it. Charge it one way
and all the light passing through it is set to "vertical" polarity, charge it another way and the light
polarity is set to "horizontal". This feature allows the pixels to be created. Each tiny cell of liquid
crystal is a pixel.
TFT (or Thin Film Transistor) is the device within each pixel that sets the charge. And so
sometimes they are called "Liquid Crystal Display" referring to the material they use or they are
called "TFT displays" referring to the tiny transistors that make them work.
The CRT works in the same way as a television - it contains an electron gun at the back of the
glass tube. This fires electrons at groups of phosphor dots which coat the inside of the screen.
When the electrons strike the phosphor dots they glow to give the colors.
Advantages of monitors
• Relatively cheap
• Reliable
• As each task is processed, the results can be displayed immediately on the screen
• Quiet
Disadvantages of monitors:
3.2.5.2 Printers
Printers are output devices. They are dedicated to creating paper copies from the computer.
Printers can produce text and images on paper. Paper can be either separate sheets such as A4 A5
A3 etc. or they may be able to print on continuous
(fanfold) paper that feed through the machine. A ream of A4 paper continuous paper with holes
on the edges, used by dot matrix printers. After you print on fanfold paper, you have to separate
the pages and tear off the edge strips. Very specialist printers can also print on plastic or even
textiles such as T shirts.
Some printers are dedicated to only producing black and white output. Their advantage is that
they are often faster than a color printer because effectively there is only one color to print
(Black). Color Printers are dedicated to creating text and images in full
color. Some types can even produce photographs when special paper is used.
There are three main types of printer that you need to know about. You will be expected to
understand the main differences i.e. purchase costs, running costs, quality and speed
3.2.5.3 Plotter
These are output devices that can produce high quality line diagrams on paper. They are often
used by engineering, architects and scientific organizations to draw plans, diagrams of machines
and printed circuit boards.
A plotter differs from a printer in that it draws images using a pen that can be lowered, raised and
moved across the page to form continuous lines. The electronically controlled pen is moved by
two computer controlled motors. The pen is lifted on and off the page by switching an
electromagnet on and off.
Flatbed plotters hold the paper still while the pens move.
Advantages:
• Larger sizes of paper can be used than would be found on most printers
Disadvantages:
• Plotters are slower than printers, drawing each line separately.
• Although drawings are completed to the highest quality they are not suitable for text (although
text can be produced)
• There is a limit to the amount of detail these plotters can produce, although there are plotters
which are "pen-less" the set are used for high-density drawings as may be used for printed circuit
board layout.
• In recent years, cheaper printers that can handle A3 and A2 sized paper have resulted in a
decline in the need for smaller plotters.
IN-TEXT QUESTION S
4.0 CONCLUSION
This unit focused on hardware devices for data processing. We need relevant input output
devices to enter data and to show the results of every processes that is carried out by the
computer. Your knowledge of computer hardware had given you basic skills to use relevant
devices to input, process and obtain information necessary foe decision-making in your
educational endeavor and business ventures.
5.0 SUMMARY
Data processing is any computer process that converts data into information. Data are any
facts, numbers, or text that can be processed by a computer. The patterns, associations, or
relationships among all this data can provide information. The CPU is a microprocessor - a
silicon chip - composed of tiny electrical switches called 'transistors'. The keyboard is the most
widely used input device and is used to enter data or commands to the computer. A Joystick is
similar to a tracker ball in operation except you have a stick which is moved rather than a rolling
ball. Graphics tablets are often used by graphics designers and illustrators. The most common
output devices are computer monitors and printers. Meta data is data about the data itself, such
as logical database design or data dictionary definitions. Resolution of a digital camera range
from about 3 million (or Mega) pixels up to 12 Mega pixels.
6.0 TUTOR-MARKED ASSIGNMENT
1. Explain what is meant by the term input device? Give three examples of input devices.
Also give possible advantages and disadvantage of the same.
2. Explain what is meant by the term output device? Give three examples of output devices.
Also give possible advantages and disadvantage of the same.
3. What are different types of printers? How a plotter is different from a printer?
Clifton, H.D. (1998). Business Data Systems: A Practical Guide to Systems Analysis and Data
Processing. London: Prentice Hill
UNIT 4: SOFTWARE ELEMENTS IN DATA PROCESSING
CONTENTS
1.0 Introduction
2.0 Objectives
4.0 Conclusion
5.0 Summary
1.0 INTRODUCTION
You have learnt from the previous unit that hardware are equipment used for data processing. They
operate in form of input, processing and output units. The equipment are useless unless some other
materials are installed into them to make them work. In this unit therefore, you will learn about software
that drive the hardware to function.
2.0 OBJECTIVES
Define software
Differentiate between system software and application software
List the functions of operating system
Enumerate the considerations to be made in selecting application software
Software is suite programs that make the hardware components of computers to function.
It is a collection of programs. Programs are computer instructions which control the works of
computers. These are computer languages which guides and directs how computers are put into
use. Software are categorized into two broad ways: System software and Application software.
Let us look at each of them one by one.
Operating system are computer-oriented software which controls the internal operation
of computer system. An operating system consists of suites of programs , one of which the
master, kernel or executive program, remains resident in the main store of the computer. This
program controls the other operating system programs in the suite and between them they
control the application programs. Example of current operating system that are commonly used
today is Windows coming up in various versions.
Priority assignment: Operating system manipulates jobs awaiting execution are scheduled
according to either a predetermined or dynamic assignment plan.
Communication: It controls data transmission between terminals and computer, and computer-
to-computer.
Software control: Software such as assemblers, compilers, utility software and subroutines are
controlled so that they available immediately when required.
Operations log: This involves maintenance of details of all jobs carried out by the computer
Application software comprises the programs that are written specifically to achieve results
appertaining to the company’s activities. In other words application software is user-oriented as
opposed to systems software which computer-oriented. Application software comes from two
sources, i.e. the company’s own DP staff or from external agencies. In the early years of DP,
purchased software was often found to be too restrictive and badly documented so that apart
from payroll was not accepted. Existing application packages are available for use. A business
application package is a complete suite of programs together with associated documentation. It
covers a business routine, is supplied by a computer manufacturer or software house, on lease
or purchase. A package is normally intended to meet the needs of a number of different user
companies. It also contains a number of options, these are selected by user by the insertion of
parameters before use.
The following benefits should be accrued from the adoption of an application package:
Portability from the existing computer system to any new computer adopted by the user.
Before any organization considers the use of application package, it is necessary to give
consideration to:
Definition of Requirements
The user company cannot abandon the study of company objectives, systems investigation and
consequent definition of DP requirements. In these respects the approach is the same as when
designing a DP systems for in-house programming.
Study of Range
A range of packages should be examined in depth before a choice is made, and existing users of
the packages should be consulted for their practical experiences and opinions. It should be
remembered that the more commonly used packages result in pressure on the
Interfacing
Consideration should be given to how a package interface with the user’s own routines, both
existing and future especially in the area of database.
Amendments
At least one member of the user company staff must be completely conversant with the
detailed operations and capabilities of the package adopted. The package supplier must be in a
position to supply all the necessary on-going amendments, especially those required through
legislation such as for taxation.
Performance
How efficient is the package in terms of its average and maximum run times on the computer?
What resources does it demand such as peripherals and of main and backing storage.
Contract Terms
The terms of contract should embrace factors such as the terms of payment, suppliers’
assistance with implementation, extent of documentation and future maintenance.
Word processing is an enhanced method of typing which converts text into form the
typewritten documents but also stores it in computer storing devices like hard disks, floppy
drives, flash drives, CD-ROM and tapes (Opateye, 1999). It involves the composition,
recording, transcription, editing and communication of text. A feature of Word processing is its
ability to interface with other text that automatically incorporates the relevant names,
descriptions and numerical data. Word processing used in offices can be categorized into the
following groups:
Replicated texts, such as standard letters and certain legal documents, perhaps with some degree
of individuality.
Updatable texts, such as reports and manufacturers’ manuals that may need amending at regular
and frequent intervals.
Be substantially non-redundant
Be program independent
Database management system therefore could be defined as a system that organizes the storage
of data in such a way to facilitate its retrieval for many different applications. The characteristics
of logical data are decided by the user’s need for flexibility in processing data. In many business
situations there is an unlimited range of demands for data; new problems and consequent new
arrangements of data occur frequently. Examples of facilities of DBMS are
Logging of transactions
Control of Passwords
Validation of data
Dynamic creation and maintenance of data dictionary
DBMS application packages available for use today are Oracles, Foxpro, Micrososft Access etc.
This involve a wide range of selection of software that is available for business
applications such as payroll, sales, purchases and nominal ledgers, stock control, financial
modeling, survey and so on.
Reflect on what you have studied in this unit by answering the following questions
4.0 CONCLUSION
Software play important roles in the operations of computer. None of the hardware
components of the computer could efficiently utilized without software that enable them to work.
Software could be system in nature which are machine language or operating system that
controls all operations and application packages that are user-oriented b ased on the demands of
activities of the user. Certain considerations should be put in place before purchasing or
developing application software.
5.0 SUMMARY
In unit 4, we have looked at software as suite programs that make the hardware
components of computers to function. It is a collection of programs. Programs are computer
instructions which control the works of computers. Operating system are computer-oriented
software which controls the internal operation of computer system. application software is user-
oriented as opposed to systems software which computer-oriented.
6.0 TUTOR-MARKED ASSIGNMENT
i. Mention five functions of Operating system.
Opateye, J.A. (1999). Elementary Computer Studies for Schools. Lagos. Arise Publications.
1.0 Introduction
2.0 Objective
4.0 Conclusion
5.0 Summary
1.0 INTRODUCTION
A computer file is a piece of arbitrary information, or resource for storing information, that is
available to a computer program and is usually based on some kind of durable storage. A file is
durable in the sense that it remains available for programs to use after the current program has
finished. Computer files can be considered as the modern counterpart of the files of printed
documents that traditionally existed in offices and libraries.
2.0 OBJECTIVES
At the conclusion of this unit you should be able to
3. CONTENTS
3.1 Concept and Contents of Files
As far as the operating system is concerned, a file is in most cases just a sequence of binary
digits. At a higher level, where the content of the file is being considered, these binary digits may
represent integer values or text characters, It is up to the program using the file to understand the
meaning and internal layout of information in the file and present it to a user as a document,
image, song, or program.
At any instant in time, a file has might have a size, normally expressed in bytes, that indicates
how much storage is associated with the file. Information in a computer file can consist of
smaller packets of information (often called records or lines) that are individually different but
share some trait in common. For example, a payroll file might contain information concerning all
the employees in a company and their payroll details; each record in the payroll file concerns just
one employee, and all the records have the common trait of being related to payroll—this is very
similar to placing all payroll information into a specific filing cabinet in an office that does not
have a computer. A text file may contain lines of text, corresponding to printed lines on a piece
of paper.
The way information is grouped into a file is entirely up to the person designing the file. This has
led to a plethora of more or less standardized file structures for all imaginable purposes, from the
simplest to the most complex. Most computer files are used by computer programs. These
programs create, modify and delete files for their own use on an as-needed basis. The
programmers who create the programs decide what files are needed, how they are to be used and
(often) their names.
In some cases, computer programs manipulate files that are made visible to the computer user.
For example, in a word-processing program, the user manipulates document files that she names
herself. The content of the document file is arranged in a way that the word-processing program
understands, but the user chooses the name and location of the file, and she provides the bulk of
the information (such as words and text) that will be stored in the file.
Files on a computer can be created, moved, modified, grown, shrunk and deleted. In most cases,
computer programs that are executed on the computer handle these operations, but the user of a
computer can also manipulate files if necessary. For instance, Microsoft Word files are normally
created and modified by the Microsoft Word program in response to user commands, but the user
can also move, rename, or delete these files directly by using a file manager program such as
Windows Explorer (on Windows computers).
Access to records in a Sequential file is serial. To reach a particular record, all the preceding
records must be read. As we observed when the topic was introduced earlier in the course, the
organization of an unordered Sequential file means it is only practical to read records from the
file and add records to the end of the file (OPEN.EXTEND).
It is not practical to delete or update records. While it is possible to delete, update and insert
records in an ordered Sequential file, these operations have some drawbacks.
Records in an ordered Sequential file are arranged, in order, on some key field or fields. When
we want to insert, delete or amend a record we must preserve the ordering. The only way to do
this is to create a new file. In the case of an insertion or update, the new file will contain the
inserted or updated record. In the case of a deletion, the deleted record will be missing from the
new file.
The main drawback to inserting, deleting or amending records in an ordered Sequential file is
that the entire file must be read and then the records written to a new file. Since disk access is
one of the slowest things we can do in computing this is very wasteful of computer time when
only a few records are involved.
For instance, if 10 records are to be inserted into a 10,000 record file, then 10,000 records will
have to be read from the old file and 10,010 written to the new file. The average time to insert a
new record will thus be very great.
1. All the records with a key value less than the record to be inserted must be read and then
written to the new file.
1. All the records with a key value less than the record to be deleted must be written to the
new file.
2. When the record to be deleted is encountered it is not written to the new file.
3. Finally, all the remaining records must be written to the new file.
1. All the records with a key value less than the record to be amended must be read and then
written to the new file.
2. Then the record to be amended must be read the amendments applied to it and the
amended record must then be written to the new file.
3. Finally, all the remaining records must be written to the new file.
As we have already noted, the problem with Sequential files is that access to the records is serial.
To reach a particular record, all the proceeding records must be read. Direct access files allow
direct access to a particular record in the file using a key and this greatly facilitates the
operations of reading, deleting, updating and inserting records. COBOL supports two kinds of
direct access file organizations -Relative and Indexed.
A Relative file may be visualized as a one dimension table stored on disk, where the Relative
Record Number is the index into the table. Relative files support sequential access by allowing
the active records to be read one after another. Relative files support only one key. The key must
be numeric and must take a value between 1 and the current highest Relative Record Number.
Enough room is allocated to the file to contain records with Relative Record Numbers between 1
and the highest record number. For instance, if the highest relative record number used is 10,000
then room for 10,000 records is allocated to the file. In this example, enough room has been
allocated on disk for 328 records. But although there is room for 328 records in the current
allocation, not all the record locations contain records. The record areas labeled "free", have not
yet had record values written to them.
To access a record in a Relative file a Relative Record Number must be provided. Supplying this
number allows the record to be accessed directly because the system can use the start position
of the file on disk, the size of the record, and the Relative Record Number to calculate the
position of the record. Because the file management system only has to make a few calculations
to find the record position the Relative file organization is the fastest of the two direct access file
organizations available in COBOL. It is also the most storage efficient.
While the usefulness of a Relative file is constrained by its restrictive key, Indexed files suffer
from no such limitation. Indexed files may have up to 255 keys, the keys can be alphanumeric
and only the primary key must be unique. In addition, it is possible to read an Indexed file
sequentially on any of its keys.
An Indexed file may have multiple keys. The key upon which the data records are ordered is
called the primary key. The other keys are called alternate keys. Records in the Indexed file are
sequenced on ascending primary key. Over the actual data records, the file system builds an
index. When direct access is required, the file system uses this index to find, read, insert, update
or delete, the required record. For each of the alternate keys specified in an Indexed file, an
alternate index is built. However, the lowest level of an alternate index does not contain actual
data records. Instead, this level made up of base records which contain only the alternate key
value and a pointer to where the actual record is. These base records are organized in ascending
alternate key order.
As well as allowing direct access to records on the primary key or any of the 254 alternate keys,
indexed files may also be processed sequentially. When processed sequentially, the records may
be read in ascending order on the primary key or on any of the alternate keys. Since the data
records are in held in ascending primary key sequence it is easy to see how the file may be
accessed sequentially on the primary key. It is not quite so obvious how sequential on the
alternate keys is achieved. This is covered in the unit on Indexed files.
Files and folders arranged in a hierarchy In modern computer systems, files are typically accessed
using names. In some operating systems, the name is associated with the file itself. In others, the
file is anonymous, and is pointed to by links that have names. In the latter case, a user can
identify the name of the link with the file itself, but this is a false analogue, especially where
there exists more than one link to the same file.
Files (or links to files) can be located in directories. However, more generally, a directory can
contain either a list of files, or a list of links to files. Within this definition, it is of paramount
importance that the term "file" includes directories. This permits the existence of directory
hierarchies. A name that refers to a file within a directory must be unique. In other words, there
must be no identical names in a directory. However, in some operating systems, a name may
include a specification of type that means a directory can contain an identical name to more than
one type of object such as a directory and a file.
In environments in which a file is named, a file's name and the path to the file's directory must
uniquely identify it among all other files in the computer system—no two files can have the same
name and path. Where a file is anonymous, named references to it will exist within a namespace.
In most cases, any name within the namespace will refer to exactly zero or one file. However,
any file may be represented within any namespace by zero, one or more names. Any string of
characters may or may not be a well-formed name for a file or a link depending upon the context
of application. Whether or not a name is well formed depends on the type of computer system
being used. Early computers permitted only a few letters or digits in the name of a file, but
modern computers allow long names (some up to 255) containing almost any combination of
unicode letters or unicode digits, making it easier to understand the purpose of a file at a glance.
Some computer systems allow file names to contain spaces; others do not. Such characters such
as / or \ are forbidden. Case-sensitivity of file names is determined by the file system.
Most computers organize files into hierarchies using folders, directories, or catalogs. Each folder
can contain an arbitrary number of files, and it can also contain other folders. These other folders
are referred to as subfolders. Subfolders can contain still more files and folders and so on, thus
building a tree-like structure in which one “master folder” (or “root folder” — the name varies
from one operating system to another) can contain any number of levels of other folders and
files. Folders can be named just as files can (except for the root folder, which often does not have
a name). The use of folders makes it easier to organize files in a logical way.
Many modern computer systems provide methods for protecting files against accidental and
deliberate damage. Computers that allow for multiple users implement file permissions to control
who may or may not modify, delete, or create files and folders. A given user may be granted
only permission to modify a file or folder, but not to delete it; or a user may be given permission
to create files or folders, but not to delete them. Permissions may also be used to allow only
certain users to see the contents of a file or folder. Permissions protect against unauthorized
tampering or destruction of information in files, and keep private information confidential by
preventing unauthorized users from seeing certain files. Another protection mechanism
implemented in many computers is a read-only flag. When this flag is turned on for a file (which
can be accomplished by a computer program or by a human user), the file can be examined, but
it cannot be modified. This flag is useful for critical information that must not be modified or
erased, such as special files that are used only by internal parts of the computer system. Some
systems also include a hidden flag to make certain files invisible; this flag is used by the
computer system to hide essential system files that users must never modify
In physical terms, most computer files are stored on hard disks—spinning magnetic disks inside
a computer that can record information indefinitely. Hard disks allow almost instant access to
computer files. On large computers, some computer files may be stored on magnetic tape. Files
can also be stored on other media in some cases, such as writeable compact discs, Zip drives,
etc.
When computer files contain information that is extremely important, a backup process is used to
protect against disasters that might destroy the files. Backing up files simply means making
copies of the files in a separate location so that they can be restored if something happens to the
computer, or if they are deleted accidentally. There are many ways to back up files. Most
computer systems provide utility programs to assist in the back-up process, which can become
very time consuming if there are many files to safeguard. Files are often copied to removable
media such as writeable CDs or cartridge tapes. Copying files to another hard disk in the same
computer protects against failure of one disk, but if it is necessary to protect against failure or
destruction of the entire computer, then copies of the files must be made on other media that can
be taken away from the computer and stored in a safe, distant location.
There are a few terms that you need to understand when learning about file system. These will be
explained over the next couple of pages. File can store data or information in various formats.
Suppose in a file data is stored in the tables just like the one below:
3.4 Records
As you saw previously, each table stores can hold a great deal of data. Each table contains a lot
of records. A record is all of the data or information about one person or one thing.
3.5 Fields
Each table contains a lot of records. A record is made up of lots of individual pieces of
information. Look at Wonder Woman's record; it stores her first name, last name, address, city
and age. Each of these individual pieces of information in a record are called a 'field'. A 'field' is
one piece of data or information about a person or thing.
What fields do you think would be stored in your student record on the school database?
IN – TEXT QUESTIONS
4.0 CONCLUSION
You have learnt of different files, their operations and how they organized in order to use
them effectively. We have relative and sequential files based on the ease of accessibility of
records store in them. Certain operations are carried out on files in form of opening, reading,
updating, committing and closing. Most computers organize files into hierarchies using folders,
directories, or catalogs. Each folder can contain an arbitrary number of files, and it can also
contain other folders. These other folders are referred to as subfolders. Subfolders can contain
still more files and folders and so on, thus building a tree-like structure in which one “master
folder” (or “root folder” — the name varies from one operating system to another) can contain
any number of levels of other folders and files. You could save your assignment and project data
in different files for you to make use of them.
5.0 SUMMARY
You covered file management and organization in this unit. Files are organized
sequentially, relatively or indexed. Most computer files are stored on hard disks and also backed
up in flash drives and CDS.
6. Explain what is meant by the term back–up? Why it is important to keep the back-up
copy away from the computer system?
7. Validation and Verification help to reduce the errors when inputting data. Justify the
statement.
8. Explain the difference between validation and verification. Give the names of three
validations checks that can be used.
CONTENTS
1.0 Introduction
2.0 Objectives
4.0 Conclusion
5.0 Summary
1.0 INTRODUCTION
Research comprises creative work undertaken on a systematic basis in order to increase the
stock of knowledge, including knowledge of man, culture and society, and the use of this stock
of knowledge to devise new applications. Research problems are questions that indicate gaps in
the scope or the certainty of our knowledge. They point either to problematic phenomena,
observed events that are puzzling in terms of our currently accepted ideas, or to problematic
theories, current ideas that are challenged by new hypotheses.
This unit, the reader will understand the characteristics of a research problem or phenomenon.
2.0 OBJECTIVES
iv. explain both internal and external criteria for research problem formulation
3.0 CONTENTS
experimentally. It is the focus or reason for engaging in our research. It is typically a topic,
phenomenon, or challenge that we are interested in and with which we are at least somewhat
familiar applications. A research problem, or phenomenon as it might be called in many forms of
qualitative research, is the topic you would like to address, investigate, or study, whether
descriptively or experimentally. It is the focus or reason for engaging in your research. It is
typically a topic, phenomenon, or challenge that you are interested in and with which you are at
least somewhat familiar.
Problem formulation is the logical first step toward this goal. As Northrop (1966) writes,
“Inquiry starts only when something is unsatisfactory, when traditional beliefs are inadequate or
in question, when the facts necessary, to resolve one’s uncertainties are not known, when the
most likely relevant hypotheses are not even imagined. What one has at the beginning of the
inquiry is merely the problem.
The formulation of research problems also has an important social function. As Merton, Broom,
and Cottrell (1959) suggest, researchers must justify the demands for attention and other scarce
resources that research makes: “In conferring upon the scientist the right to claim that a question
deserves the concerted attention of others as well as himself, the social institution of science
exacts the obligation that he justify the claim.
Achieving significant research results is perhaps the most powerful justification for such claims,
but this type of justification can be offered only after the fact, and only in the event that the
research is successful. A compelling research problem, by contrast, must marshal support in
advance of research and, if it is sufficiently compelling, can even sustain that support through
the sometimes fruitless periods that researcher's experience. However, despite research
problems’ logical priority in inquiry, and their importance as a priori justification, a problem
formulation, as John Dewey stresses, is in fact a “progressive” matter.
1. On the one hand, new theories solve research problems by Research Problem: accounting
for unexplained phenomena and by superseding questionable older theories.
2. On the other hand, existing theory guides researchers in formulating research problems.
In determining whether and in what respects a phenomenon or a theory is problematic,
researchers consider the context of accumulated theoretical as well as empirical
knowledge.
Daily problems
Technological changes
Unexplored areas
The selection of one appropriate researchable problem out of the identified problems requires
evaluation of those alternatives against certain criteria, which may be grouped into:
1. Researcher’s interest: The problem should interest the researcher and be a challenge to
him. Without interest and curiosity, he may not develop sustained perseverance Interest
in a problem depends upon the researcher’s educational background, experience, outlook
and sensitivity.
2. Researcher’s own resource: In the case of a research to be done by a researcher on his
own, consideration of his own financial resource is pertinent. If it is beyond his means, he
will not be able to complete the work, unless he gets some external financial support.
Time resource is more important than finance. Research is a time consuming process;
hence it should be properly utilized.
3. Researcher’s competence: A mere interest in a problem will not do. The researcher must
be competent to plan and carry out a study of the problem. He must possess adequate
knowledge of the subject-matter, relevant methodology and statistical
procedures.
1. Research-ability of the problem: The problem should be researchable, i.e., amendable for
finding answers to the questions involved in it through the scientific method.
2. Novelty of the problem: The problem must have novelty. There is no use of wasting
one’s time and energy on a problem already studied thoroughly by others.
3. Importance and urgency: Problems requiring investigation are unlimited, but available
research efforts are very much limited.
4. Facilities: Research requires certain facilities such, as well equipped library facility,
suitable and competent guidance, data analysis facility, etc. Hence the availability of the
facilities relevant to the problem must be considered. Problems for research, their relative
importance and significance should be considered.
5. Feasibility: A problem may be a new one and also important, but if research on it is not
feasible, it cannot be selected.
6. Usefulness and social relevance: Above all, the study of the problem should make a
These help to ensure that your study will remain manageable and that you will remain motivated.
1. Interest: a research endeavour is usually time consuming, and involves hard work and possibly
unforeseen problems. One should select topic of great interest to sustain the required motivation.
2. Magnitude: It is extremely important to select a topic that you can manage within the time and
resources at your disposal. Narrow the topic down to something manageable, specific and clear.
3. Level of expertise: Make sure that you have an adequate level of expertise for the task you are
proposing since you need to do the work yourself.
4. Relevance: Ensure that your study adds to the existing body of knowledge, bridges
current gaps and is useful in policy formulation. This will help you to sustain interest in the
study.
5. Availability of data: Before finalizing the topic, make sure that data are available.
6. Ethical issues: How ethical issues can affect the study and how ethical problemscan be
overcome should be thoroughly examined at the problem formulating stage.
The rational- theoretical or practical (why one wants to have the questions answered?)
The specifying questions (possible answers to the originating) questions in term of that satisfy
the rationale.)
Are formulated in such specific indicate where exactly the answers to them can be
searched for.
May be formulated in terms of broader delimited categories of social variable but do not
indicate specifically which particular variables in each class might be germane to the
issues.
Usually derive from a general theoretical orientation rather than a definite theory.
Rationale of Questions
practice.
Specifying Questions
Involve the breaking down of originating question in with several specifying questions
related to particular aspects and their consequences.
The following are detailed list of criteria for the choice of research problem.
Novelty-It should be sufficiently original so that it does not involve objectionable duplication.
Importance-If it is not worth-while, if adds to neither knowledge nor lead to any improvements
in the current practices.
Immediate Application-The investigator should ask himself the question, will my research help
in solving an urgent problem
Availability of data
Availability of cooperation
Availability of guidance
- Naming a Broad Filed to choose the broad area of study instead of specific problem
makes no justification.
- Narrowing or Localizing a Topic
- The problem should not be narrowed to such an extent that it becomes too small and
insignificant from research point or view.
IN-TEXT QUESTIONS
Now that you have finished studying this unit, attempt these questions to revise what you
have learnt
4.0 CONCLUSION
In concluding lines we need to say that research is a creative work and this unit gave the
researcher an outline about the identification and formulation of research problems. Best selected
problems serve its purpose and prove the solution to many identified and unidentified problems.
5.0 SUMMARY
This unit exposed you to intricacies involved in research problem formulation. We also
looked at both internal and external criteria for generating research problem. Sources for
research problem includes schools, literatures, experiences, Special programme, Research
reports, term papers, consultation with, course instructor, advisor, faculty member among others.
i. Explain both internal and external criteria for research problem formulation
ii. Discuss considerations in selecting a research problem
iii. Discuss characteristics of a good research problem.
CONTENTS
1.0 Introduction
2.0 Objectives
3.0 Main Content
3.1 Definition of Research Question
3.2 Characteristics of Research Question
3.3 Concept and Types of Hypotheses
3.4 Types of Hypotheses
3.5 Characteristics of Hypotheses
3.6 Functions of Hypotheses
4.0 Conclusion
5.0 Summary
6.0 Tutor-Marked Assignment
7.0 References / Further Readings
1.0 INTRODUCTION
Long before you create a research proposal, let alone conduct your research, you need to identify a
problem to address and then a question or questions to ask regarding your targeted problem. This chapter
first discusses the nature of a research problem, where you might get ideas for a problem to investigate,
narrowing down or focusing on a particular problem to address, and writing good research questions. It
then discusses finding literature that is relevant to and helpful in clarifying your targeted problem and
question(s).
2.0 OBJECTIVES
For research question to be complete and embracing it has to have the following features:
FINER
Feasible
Adequate numbers of subjects?
Adequate technical expertise?
Affordable in time and money?
Is it possible to measure or manipulate the variables?
Interesting to the investigator?
Novel to the field?
ETHICAL
Potential harm to subjects?
Potential breech of subject confidentiality?
- Relevant to scientific knowledge/theory, organizational, health or social management and
policy and to individual welfare.
1. Be sure to read on the topic to familiarize yourself with it before making a final decision.
You need to make certain that the topic is researchable in an interdisciplinary sense,
meaning that there is sufficient published material on the topic in the legal literature and
in the literature of a second or possibly a third discipline to write a 35-page paper.
2. As noted, a research hypothesis is more than just a topic. It has two elements (variables)
that are in relation to each other. Remember that, within the word "hypothesis" is the
word "thesis." Your hypothesis is what you propose to “prove” by your research. As a
result of your research, you will arrive at a conclusion, a theory, or understanding that
will be useful or applicable beyond the research itself.
3. Avoid judgmental words in your hypothesis. Value judgments are subjective and are not
appropriate for a hypothesis. You should strive to be objective. Therefore the use of
personal opinion is to be avoided.
4. Your hypothesis must involve an issue or question that cannot be answered exclusively
by the discipline of law. You should try to limit your inquiry to the literatures of 2 or 3
disciplines. It is best to choose a hypothesis where you already have some level of
familiarity with the disciplines that are most relevant to the topic.
5. Be sure that each term in your hypothesis is clearly understood and defined; do not deal
in generalities or assume that the reader knows the meaning of a technical term.
6. Specify, if appropriate, whether you will be dealing with state or federal law or both on a
comparative basis if appropriate.
7. Know that your hypothesis may change over time as your research progresses. You must
obtain the professor's approval of your hypothesis, as well as any modifications to your
hypothesis, before proceeding with any work on the topic.
Your will be expressing your hypothesis in 3 ways:
As a one-sentence hypothesis
As a research question
As a title for your paper
Your hypothesis will become part of your research proposal.
Note how each student, in the samples below, began with a general topic or area of interest,
which evolved into a hypothesis. Look for the variables that each student subsequently explored
in his/her paper. The examples below are final form hypotheses, which have been revised
throughout the research process. You will find that your hypothesis may undergo changes too, as
your research progresses.
Examples:
i. There is no significant difference in achievement between male and female Business
Education students
ii. There is no significant relationship between students’ attitudes and academic
performance in research methods
iii. There is no significant main effect of socio-economic background of students’ motivation
to mathematics
3.4 TYPES OF HYPOTHESES
There are two types of hypotheses: Null hypothesis and Alternate hypothesis.
3.4.1 Null Hypothesis
This deals with no difference between parameters being measured. The following are examples
of null hypotheses:
i. There is no significant difference in attitudes towards Business Education between male
and female distance education students
ii. There is no significant relationship between students’ reading habit and academic
performance in Research methods
iii. There is no significant main effect of employment commitment on ODL students’ class
of degree after convocation.
Null hypothesis is denoted by Ho
3.4.2 Alternative Hypothesis
Alternative hypothesis states categorically the direction of difference. It is therefore referred to
as directional hypothesis. Examples are
Attitudes of female distance learning students to Business Education is more favourable than
their male counterparts
The relationship between students’ reading habit and academic performance is negative
There is no significant main effect of employment commitment on ODL students’ class of
degree after convocation.
Null hypothesis is denoted by H1.
You need to note that in educational research, hypotheses are tested at 0.05 level of significant.
IN-TEXT QUESTIONS
You can now reflect on what you have studied
A research question should be debatable and of interest to both you and your potential
readers. True / False
Distinguish between research question and hypothesis
Mention two types of hypothesis
4.0 CONCLUSION
After you have identified your research problem, the next step is to derive your research
questions and generate hypotheses. This unit had explained what research questions are with
various considerations you need to take to form them. Also, hypotheses are generated in form of
null or alternative. Hypotheses are met to be plausible, testable and not ambiguous.
5.0 SUMMARY
Research questions are not necessarily an attempt to answer the many philosophical questions
that often arise in schools, and they are certainly not intended to be an avenue for grinding personal axes
regarding classroom or school issues. The research question flows from the topic that you are
considering. The research question, when stated as one sentence, is your Research Hypothesis.
Null and alternative hypotheses existed but in educational research it is advisable to state
hypotheses in null form.
1.0 Introduction
2.0 Objectives
3.0 Main Content
3.1 Research Approaches
3.1.1 Quantitative Research
3.1.2 Qualitative Research
3.1.3 Advocacy / Participatory Approach
3.2 Research Designs
3.2.1 Action Research Designs
3.2.2 Case Study Design
3.2.3 Causal-Comparative Design
3.2.4 Cohort Design
3.2.5 Cross-sectional Design
4.0 Conclusion
5.0 Summary
6.0 Tutor-Marked Assignment
7.0 References / Further Readings
The process
Researchers will have one or more hypotheses. These are the questions that they want to address
which include predictions about possible relationships between the things they want to
investigate (variables). In order to find answers to these questions, the researchers will also have
various instruments and materials (e.g. paper or computer tests, observation check lists etc.) and
a clearly defined plan of action. Data is collected by various means following a strict procedure
and prepared for statistical analysis. Nowadays, this is carried out with the aid of sophisticated
statistical computer packages. The analysis enables the researchers to determine to what extent
there is a relationship between two or more variables. This could be a simple association (e.g.
people who exercise on a daily basis have lower blood pressure) or a causal relationship (e.g.
daily exercise actually leads to lower blood pressure). Statistical analysis permits researchers to
discover complex causal relationships and to determine to what extent one variable influences
another.
The results of statistical analyses are presented in journals in a standard way, the end result being
a P value. For people who are not familiar with scientific research jargon, the discussion sections
at the end of articles in peer reviewed journals usually describe the results of the study and
explain the implications of the findings in straightforward terms
Principles
Objectivity is very important in quantitative research. Consequently, researchers take great care
to avoid their own presence, behaviour or attitude affecting the results (e.g. by changing the
situation being studied or causing participants to behave differently). They also critically
examine their methods and conclusions for any possible bias.
Researchers go to great lengths to ensure that they are really measuring what they claim to be
measuring. For example, if the study is about whether background music has a positive impact
on restlessness in residents in a nursing home, the researchers must be clear about what kind of
music to include, the volume of the music, what they mean by restlessness, how to measure
restlessness and what is considered a positive impact. This must all be considered, prepared and
controlled in advance.
External factors, which might affect the results, must also be controlled for. In the above
example, it would be important to make sure that the introduction of the music was not
accompanied by other changes (e.g. the person who brings the CD player chatting with the
residents after the music session) as it might be the other factor which produces the results (i.e.
the social contact and not the music). Some possible contributing factors cannot always be ruled
out but should be acknowledged by the researchers.
The main emphasis of quantitative research is on deductive reasoning which tends to move from
the general to the specific. This is sometimes referred to as a top down approach. The validity of
conclusions is shown to be dependent on one or more premises (prior statements, findings or
conditions) being valid. Aristotle’s famous example of deductive reasoning was: All men are
mortal àSocrates is a man à Socrates is mortal. If the premises of an argument are inaccurate,
then the argument is inaccurate. This type of reasoning is often also associated with the fictitious
character Sherlock Holmes. However, most studies also include an element of inductive
reasoning at some stage of the research (see section on qualitative research for more details).
Researchers rarely have access to all the members of a particular group (e.g. all people with
dementia, carers or healthcare professionals). However, they are usually interested in being able
to make inferences from their study about these larger groups. For this reason, it is important that
the people involved in the study are a representative sample of the wider population/group.
However, the extent to which generalizations are possible depends to a certain extent on the
number of people involved in the study, how they were selected and whether they are
representative of the wider group. For example, generalizations about psychiatrists should be
based on a study involving psychiatrists and not one based on psychology students. In most
cases, random samples are preferred (so that each potential participant has an equal chance of
participating) but sometimes researchers might want to ensure that they include a certain number
of people with specific characteristics and this would not be possible using random sampling
methods. Generalizability of the results is not limited to groups of people but also to situations. It
is presumed that the results of a laboratory experiment reflect the real life situation which the
study seeks to clarify.
When looking at results, the P value is important. P stands for probability. It measures the
likelihood that a particular finding or observed difference is due to chance. The P value is
between 0 and 1. The closer the result is to 0, the less likely it is that the observed difference is
due to chance. The closer the result is to 1, the greater the likelihood that the finding is due to
chance (random variation) and that there is no difference between the groups/variables.
Qualitative research is the approach usually associated with the social constructivist paradigm
which emphasises the socially constructed nature of reality. It is about recording, analysing and
attempting to uncover the deeper meaning and significance of human behaviour and experience,
including contradictory beliefs, behaviours and emotions. Researchers are interested in gaining a
rich and complex understanding of people’s experience and not in obtaining information which
can be generalized to other larger groups.
The process
The approach adopted by qualitative researchers tends to be inductive which means that they
develop a theory or look for a pattern of meaning on the basis of the data that they have
collected. This involves a move from the specific to the general and is sometimes called a
bottom-up approach. However, most research projects also involve a certain degree of deductive
reasoning (see section on quantitative research for more details).
Data collection may be carried out in several stages rather than once and for all. The researchers
may even adapt the process mid-way, deciding to address additional issues or dropping questions
which are not appropriate on the basis of what they learn during the process. In some cases, the
researchers will interview or observe a set number of people. In other cases, the process of data
collection and analysis may continue until the researchers find that no new issues are emerging.
Principles
Researchers will tend to use methods which give participants a certain degree of freedom and
permit spontaneity rather than forcing them to select from a set of pre-determined responses (of
which none might be appropriate or accurately describe the participant’s thoughts, feelings,
attitudes or behaviour) and to try to create the right atmosphere to enable people to express
themselves. This may mean adopting a less formal and less rigid approach than that used in
quantitative research.
It is believed that people are constantly trying to attribute meaning to their experience. Therefore,
it would make no sense to limit the study to the researcher’s view or understanding of the
situation and expect to learn something new about the experience of the participants.
Consequently, the methods used may be more open-ended, less narrow and more exploratory
(particularly when very little is known about a particular subject). The researchers are free to go
beyond the initial response that the participant gives and to ask why, how, in what way etc. In
this way, subsequent questions can be tailored to the responses just given. Qualitative research
often involves a smaller number of participants. This may be because the methods used such as
in-depth interviews are time and labour intensive but also because a large number of people are
not needed for the purposes of statistical analysis or to make generalizations from the results.
The smaller number of people typically involved in qualitative research studies and the greater
degree of flexibility does not make the study in any way “less scientific” than a typical
quantitative study involving more subjects and carried out in a much more rigid manner. The
objectives of the two types of research and their underlying philosophical assumptions are
simply different. However, as discussed in the section on “philosophies guiding research”, this
does not mean that the two approaches cannot be used in the same study.
The pragmatic approach to science involves using the method which appears best suited to the
research problem and not getting caught up in philosophical debates about which is the best
approach. Pragmatic researchers therefore grant themselves the freedom to use any of the
methods, techniques and procedures typically associated with quantitative or qualitative research.
They recognise that every method has its limitations and that the different approaches can be
complementary.
They may also use different techniques at the same time or one after the other. For example, they
might start with face-to-face interviews with several people or have a focus group and then use
the findings to construct a questionnaire to measure attitudes in a large scale sample with the aim
of carrying out statistical analysis.
Depending on which measures have been used, the data collected is analysed in the appropriate
manner. However, it is sometimes possible to transform qualitative data into quantitative data
and vice versa although transforming quantitative data into qualitative data is not very common.
Being able to mix different approaches has the advantages of enabling triangulation.
Triangulation is a common feature of mixed methods studies. It involves, for example:
In some studies, qualitative and quantitative methods are used simultaneously. In others, first one
approach is used and then the next, with the second part of the study perhaps expanding on the
results of the first. For example, a qualitative study involving in-depth interviews or focus group
discussions might serve to obtain information which will then be used to contribute towards the
development of an experimental measure or attitude scale, the results of which will be analysed
statistically.
The researchers may adopt a less neutral position than that which is usually required in scientific
research. This might involve interacting informally or even living amongst the research
participants (who are sometimes referred to as co-researchers in recognition that the study is not
simply about them but also by them). The findings of the research might be reported in more
personal terms, often using the precise words of the research participants. Whilst this type of
research could by criticised for not being objective, it should be noted that for some groups of
people or for certain situations, it is necessary as otherwise the thoughts, feelings or behaviour of
the various members of the group could not be accessed or fully understood. Vulnerable groups
are rarely in a position of power within society. For this reason, researchers are sometimes
members of the group they are studying or have something in common with the members of the
group.
3.2 Research Designs
The function of a research design is to ensure that the evidence obtained enables you to effectively
address the research problem logically and as unambiguously as possible. In social sciences research,
obtaining information relevant to the research problem generally entails specifying the type of evidence
needed to test a theory, to evaluate a program, or to accurately describe and assess meaning related to an
observable phenomenon.
With this in mind, a common mistake made by researchers is that they begin their investigations far too
early, before they have thought critically about what information is required to address the study's
research questions. Without attending to these design issues beforehand, the overall research problem will
not be adequately addressed and any conclusions drawn will risk being weak and unconvincing. As a
consequence, the overall validity of the study will be undermined.
Given this, the length and complexity of describing research designs in your paper can vary considerably,
but any well-developed design will achieve the following:
Identify the research problem clearly and justify its selection, particularly in relation to
any valid alternative designs that could have been used,
Review and synthesize previously published literature associated with the problem,
Clearly and explicitly specify hypotheses [i.e., research questions] central to the research
problem,
Effectively describe the data which will be necessary for an adequate testing of the
hypotheses and explain how such data will be obtained, and
Describe the methods of analysis to be applied to the data in determining whether or not
the hypotheses are true or false.
This is a collaborative and adaptive research design that lends itself to use in work or community
situations. Design focuses on pragmatic and solution-driven research outcomes rather than
testing theories. When practitioners use action research, it has the potential to increase the
amount they learn consciously from their experience; the action research cycle can be regarded
as a learning cycle. Action research studies often have direct and obvious relevance to
improving practice and advocating for change. There are no hidden controls or preemption of
direction by the researcher.
A case study is an in-depth study of a particular research problem rather than a sweeping statistical survey
or comprehensive comparative inquiry. It is often used to narrow down a very broad field of research into
one or a few easily researchable examples. The case study research design is also useful for testing
whether a specific theory and model actually applies to phenomena in the real world. It is a useful design
when not much is known about an issue or phenomenon.
A single or small number of cases offers little basis for establishing reliability or to
generalize the findings to a wider population of people, places, or things.
Intense exposure to the study of a case may bias a researcher's interpretation of the
findings.
Design does not facilitate assessment of cause and effect relationships.
Vital information may be missing, making the case hard to interpret.
The case may not be representative or typical of the larger problem being investigated.
If the criteria for selecting a case is because it represents a very unusual or unique
phenomenon or problem for study, then your interpretation of the findings can only apply
to that particular case.
Appropriate time order -- to conclude that causation was involved, one must see that cases were
exposed to variation in the independent variable before variation in the dependent variable.
Non-spuriousness -- a relationship between two variables that is not due to variation in a third
variable.
Causality research designs assist researchers in understanding why the world works the
way it does through the process of proving a causal link between variables and by the
process of eliminating other possibilities.
Replication is possible.
There is greater confidence the study has internal validity due to the systematic subject
selection and equity of groups being compared.
Not all relationships are casual! The possibility always exists that, by sheer coincidence,
two unrelated events appear to be related
Conclusions about causal relationships are difficult to determine due to a variety of
extraneous and confounding variables that exist in a social environment. This means
causality can only be inferred, never proven.
If two variables are correlated, the cause must come before the effect. However, even
though two variables might be causally related, it can sometimes be difficult to determine
which variable comes first and, therefore, to establish which variable is the actual cause
and which the actual effect is.
Closed Cohort Studies [static populations, such as patients entered into a clinical trial] involve
participants who enter into the study at one defining point in time and where it is presumed that
no new participants can enter the cohort. Given this, the number of study participants remains
constant (or can only decrease).
The use of cohorts is often mandatory because a randomized control study may be
unethical. For example, you cannot deliberately expose people to asbestos, you can only
study its effects on those who have already been exposed. Research that measures risk
factors often relies upon cohort designs.
Because cohort studies measure potential causes before the outcome has occurred, they
can demonstrate that these “causes” preceded the outcome, thereby avoiding the debate
as to which is the cause and which is the effect.
Cohort analysis is highly flexible and can provide insight into effects over time and
related to a variety of different types of changes [e.g., social, cultural, political, economic
etc.].
Either original data or secondary data can be used in this design.
In cases where a comparative analysis of two cohorts is made [e.g., studying the effects
of one group exposed to asbestos and one that has not], a researcher cannot control for all
other factors that might differ between the two groups. These factors are known as
confounding variables.
Cohort studies can end up taking a long time to complete if the researcher must wait for
the conditions of interest to develop within the group. This also increases the chance that
key variables change during the course of the study, potentially impacting the validity of
the findings.
Due to the lack of randomization in the cohort design, its external validity is lower than
that of study designs where the researcher randomly assigns participants.
Cross-sectional research designs have three distinctive features: no time dimension; a reliance on existing
differences rather than change following intervention; and, groups are selected based on existing
differences rather than random allocation. The cross-sectional design can only measure differences
between or from among a variety of people, subjects, or phenomena rather than a process of change. As
such, researchers using this design can only employ a relatively passive approach to making causal
inferences based on findings.
Cross-sectional studies provide a clear 'snapshot' of the outcome and the characteristics
associated with it, at a specific point in time.
Unlike an experimental design, where there is an active intervention by the researcher to
produce and measure change or to create differences, cross-sectional designs focus on
studying and drawing inferences from existing differences between people, subjects, or
phenomena.
Entails collecting data at and concerning one point in time. While longitudinal studies
involve taking multiple measures over an extended period of time, cross-sectional
research is focused on finding relationships between variables at one moment in time.
Groups identified for study are purposely selected based upon existing differences in the
sample rather than seeking random sampling.
Cross-section studies are capable of using data from a large number of subjects and,
unlike observational studies, is not geographically bound.
Can estimate prevalence of an outcome of interest because the sample is usually taken
from the whole population.
Because cross-sectional designs generally use survey techniques to gather data, they are
relatively inexpensive and take up little time to conduct.
IN-TEXT QUESTIONS
Now that you’ve been introduced to research approaches and designs in this unit, ascertain the
level of your attainment by proffer answers to the following questions
4.0 CONCLUSION
You have discovered from this unit that the main emphasis of quantitative research is on
deductive reasoning which tends to move from the general to the specific. This is sometimes
referred to as a top down approach. In qualitative research, researchers are interested in gaining
a rich and complex understanding of people’s experience and not in obtaining information which
can be generalized to other larger groups. It is also essential that without attending design issues
beforehand, the overall research problem will not be adequately addressed and any conclusions drawn
will risk being weak and unconvincing. As a consequence, the overall validity of the study will be
undermined.
5.0 SUMMARY
This unit had exposed you to three major approaches of research, namely quantitative, qualitative
and advocacy approaches. It is also discovered that research designs are essential as to be
considered when one wants to undertake a research. Action research, case study, causal
comparative, cohort and cross-sectional designs are applicable to educational and business
related studies. Consider the strengths and weaknesses of any design that you want to employ to
have a solid ground on which the work is based.
6.0 TUROR-MARKED ASSIGNMENT
Bell, J. (1999). Doing Your Research Project. Buckingham: Oxford University Press
Kerlinger, F.N. & Lee (2000) Foundations of Behavioural Research. Australia: Thomson
Learning Inc.
Nwana, O.C. (1990) Introduction to Educational Research. Ibadan: Heinemann Educational
Books (Nigeria) Limited
UNIT 4: RESEARCH APPROACHES AND DESIGNS II
CONTENTS
1.0 Introduction
2.0 Objectives
3.0 Main Contents
3.1 Research Designs
3.1.1 Descriptive Survey Design
3.1.2 Experimental Design
3.1.3 Exploratory Design
3.1.4 Historical Design
3.1.5 Longitudinal Design
3.1.6 Meta-analysis Design
3.1.7 Observational Design
4.0 Conclusion
5.0 Summary
6.0 Tutor-Marked Assignment
7.0 References / Further Readings
1.0 INTRODUCTION
2.0 OBJECTIVES
3.0 CONTENTS
The results from a descriptive research cannot be used to discover a definitive answer or
to disprove a hypothesis.
Because descriptive designs often utilize observational methods [as opposed to
quantitative methods], the results cannot be replicated.
The descriptive function of research is heavily dependent on instrumentation for
measurement and observation.
A blueprint of the procedure that enables the researcher to maintain control over all factors that
may affect the result of an experiment. In doing this, the researcher attempts to determine or predict what
may occur. Experimental research is often used where there is time priority in a causal relationship (cause
precedes effect), there is consistency in a causal relationship (a cause will always lead to the same effect),
and the magnitude of the correlation is great. The classic experimental design specifies an experimental
group and a control group. The independent variable is administered to the experimental group and not to
the control group, and both groups are measured on the same dependent variable. Subsequent
experimental designs have used more groups and more measurements over longer periods. True
experiments must have control, randomization, and manipulation.
The design is artificial, and results may not generalize well to the real world.
The artificial settings of experiments may alter the behaviors or responses of participants.
Experimental designs can be costly if special equipment or facilities are needed.
Some research problems cannot be studied using an experiment because of ethical or
technical reasons.
Difficult to apply ethnographic and other qualitative methods to experimentally designed
studies.
An exploratory design is conducted about a research problem when there are few or no earlier studies to
refer to or rely upon to predict an outcome. The focus is on gaining insights and familiarity for later
investigation or undertaken when research problems are in a preliminary stage of investigation.
Exploratory designs are often used to establish an understanding of how best to proceed in studying an
issue or what methodology would effectively apply to gathering information about the issue.
The goals of exploratory research are intended to produce the following possible insights:
Exploratory research generally utilizes small sample sizes and, thus, findings are
typically not generalizable to the population at large.
The exploratory nature of the research inhibits an ability to make definitive conclusions
about the findings. They provide insight but not definitive conclusions.
The research process underpinning exploratory studies is flexible but often unstructured,
leading to only tentative results that have limited value to decision-makers.
Design lacks rigorous standards applied to methods of data gathering and analysis
because one of the areas for exploration could be to determine what method or
methodologies could best fit the research problem.
The purpose of a historical research design is to collect, verify, and synthesize evidence from the past to
establish facts that defend or refute a hypothesis. It uses secondary sources and a variety of primary
documentary evidence, such as, diaries, official records, reports, archives, and non-textual information
[maps, pictures, audio and visual recordings]. The limitation is that the sources must be both authentic
and valid.
The historical research design is unobtrusive; the act of research does not affect the
results of the study.
The historical approach is well suited for trend analysis.
Historical records can add important contextual background required to more fully
understand and interpret a research problem.
There is often no possibility of researcher-subject interaction that could affect the
findings.
Historical sources can be used over and over to study different research problems or to
replicate a previous study.
A longitudinal study follows the same sample over time and makes repeated observations. For example,
with longitudinal surveys, the same group of people is interviewed at regular intervals, enabling
researchers to track changes over time and to relate them to variables that might explain why the changes
occur. Longitudinal research designs describe patterns of change and help establish the direction and
magnitude of causal relationships. Measurements are taken on each variable over two or more distinct
time periods. This allows the researcher to measure change in variables over time. It is a type of
observational study sometimes referred to as a panel study.
Meta-analysis is an analytical methodology designed to systematically evaluate and summarize the results
from a number of individual studies, thereby, increasing the overall sample size and the ability of the
researcher to study effects of interest. The purpose is to not simply summarize existing knowledge, but to
develop a new understanding of a research problem using synoptic reasoning. The main objectives of
meta-analysis include analyzing differences in the results among studies and increasing the precision by
which effects are estimated. A well-designed meta-analysis depends upon strict adherence to the criteria
used for selecting studies and the availability of information in each study to properly analyze their
findings. Lack of information can severely limit the type of analyses and conclusions that can be reached.
In addition, the more dissimilarity there is in the results among individual studies [heterogeneity], the
more difficult it is to justify interpretations that govern a valid synopsis of results.
- A meta-analysis needs to fulfill the following requirements to ensure the validity of your findings:
This type of research design draws a conclusion by comparing subjects against a control group, in cases
where the researcher has no control over the experiment. There are two general types of observational
designs. In direct observations, people know that you are watching them. Unobtrusive measures involve
any method for studying behavior where individuals do not know they are being observed. An
observational study allows a useful insight into a phenomenon and avoids the ethical and practical
difficulties of setting up a large and cumbersome research project.
o Observational studies are usually flexible and do not necessarily need to be structured
around a hypothesis about what you expect to observe [data is emergent rather than pre-
existing].
o The researcher is able to collect in-depth information about a particular behavior.
o Can reveal interrelationships among multifaceted dimensions of group interactions.
o You can generalize your results to real life situations.
o Observational research is useful for discovering what variables may be important before
applying other methods like experiments.
o Observation research designs account for the complexity of group behaviors.
o Reliability of data is low because seeing behaviors occur over and over again may be a
time consuming task and are difficult to replicate.
o In observational research, findings may only reflect a unique sample population and,
thus, cannot be generalized to other groups.
o There can be problems with bias as the researcher may only "see what they want to see."
o There is no possibility to determine "cause and effect" relationships since nothing is
manipulated.
o Sources or subjects may not all be equally credible.
o Any group that is knowingly studied is altered to some degree by the presence of the
researcher, therefore, potentially skewing any data collected.
IN-TEXT QUESTIONS
Make a quick revision of various research designs discussed in this unit by proffer answers to the
following questions
4.0 CONCLUSION
As you have studied in this unit, research is carried out based on the design employ by
the researcher. It is required that such researchers need to be conversant with the designs
available so as to use the appropriate ones to undertake the research. Your understanding of
research designs give you the in road to carry out your research project as a n education student.
5.0 SUMMARY
In this unit, descriptive survey design is used to obtain information concerning the current status of
the phenomena and to describe "what exists" with respect to variables or conditions in a situation. Other
research designs like experimental, meta-analysis, historical, explorative and longitudinal were discussed.
Each of the designs is effective in its own right, notwithstanding, each also has strengths and limitations
that you need to consider during research processes.
6.0 TUTOR-MARKED ASSIGNMENT
o Discuss key features of descriptive survey design
o Mention four advantages and four disadvantages of experimental research
o List five insights the goals of explorative research intended to produce
o Differentiate between historical and longitudinal designs.
7.0 REFERENCES / FURTHER READINGS
Anastas, J. W. (1999). Flexible Methods: Relational and Longitudinal Research. New York: Columbia
Cooper, H., Larry V. H., & Jeffrey C. V. (2009). The Handbook of Research Synthesis and Meta-
Analysis. New York: Russell Sage Foundation. University Press.
De Vaus, D. A. (2001). Research Design in Social Research. London: SAGE.
Gall, M. (2007). Educational Research: An Introduction. Boston, MA: Pearson/Allyn and Bacon.
Maykut, P. S.(1994). Beginning Qualitative Research: A Philosophic and Practical Guide. Washington,
D.C.: Falmer Press
Stanford Encyclopedia of Philosophy (2013). Metaphysics Research Lab, CSLI, Stanford University
Rosenbaum, P. R. (2010). Design of Observational Studies. New York: Springer.
Reason, P. & Hilary B. (2001). Handbook of Action Research: Participative Inquiry and Practice.
Thousand Oaks, CA: SAGE.
UNIT 5: Data Collection Tools and Methods
CONTENTS
1.0 Introduction
2.0 Objectives
3.0 Main Contents
3.1 Data Collection Tools
3.1.1 Anecdotal Records
3.1.2 Customer and Vendor Feedback Records
3.1.3 Interviews
3.1.4 Focus Groups
3.1.5 Tasting Panels
3.1.6 Observation
3.1.7 Document Analysis
3.1.8 Questionnaires
4.0 Conclusion
5.0 Summary
6.0 Tutor-Marked Assignment
7.0 References / Further Readings
1.0 INTRODUCTION
To grow your market, you need to assess its current performance, strengths, and weaknesses.
Assessment involves research—collecting, analyzing, and interpreting data. There are a number of
research methods and tools that can be used to collect data on the performance of your farmers
market. There are a number of different methods of collecting the data you need to assess your
market’s performance, evaluate your strengths and weaknesses and position relative to your
competitors, and create a strategic marketing plan. This section provides a brief overview of those
data collection methods and their appropriate use.
2.0 Objectives
At the end of this unit, you should be able to
3.0 CONTENTS
3.6 Observation
Another useful source of data is observation with specific criteria or questions in mind. Market
management might gather data by observing and recording the level of activity at the market at
different hours— is the market busy early in the day with business tapering off later or is it steady?
Do vendors tend to have large amounts of product left or run out early? Observation, especially
when combined with other sources of data, can be very useful in evaluating the market’s
performance. One method of observation involves counting—numbers of customers per half hour or
the number of customers who fit certain categories, such as families with young children or seniors,
for example. Especially when evaluating your position relative to competitors such as supermarkets,
you can use observations and comparisons; for example, you may compare price, quality, or selection
of a sample of produce.
Apart from business research, observation is also applicable to education. Observation of students
is an important aspect of classroom activities. It is most effective when focused on skills,
concepts, and attitudes. Making brief notes and keeping checklists when watching students’
activities during practical sections like typing enable teachers to keep records of continuous
progress and achievement. Observation helps researchers to create specific occasions for
observation and practice focus- that is, they select what and whom to observe, and put all else in
the background.
1. A carefully planned observation, in which observer knows what they are looking for
and what is relevant in a situation.
2. Observers are aware of the wholeness of what is observed.
3. Observers are objective.
4. Observers separate the facts from the interpretation of facts.
5. Observations are checked and verified, whenever possible by repetition or by
comparison with those of other competent observers.
6. Observations are carefully and expertly recorded.
3.8 Questionnaire
A questionnaire is a series of relevant questions which are usually used to elicit information
from respondent who are normally drawn from the target population of a given study. A
questionnaire may contain either structure or unstructured questions. The structure question
otherwise known as close questions are those item in a questionnaire in which alternative
responses are provided by the researcher from which the respondent is to select from. Example
of a structure item of a questionnaire
In a period of dwindling national economy, parent should be responsible for paying the school
fees of their wards at all level of education.
Alternative Responses Provided
A. Strongly agree
B. Agree
C. Undecided
D. Disagree
E. Strongly disagree.
An unstructured question or item (or open-ended item) is that item or question in which pre-
determine responses are not provided for respondent to choose from, Example of an unstructured
item of a questionnaire:
Questions: What do you consider to be the implication of open and distant learning on the future
of a national economy? ----------------------------------------------------------------------------------------
---------------------------------------------------------------------------------------------------------------------
---------------------------------------------------------------------------------------------------------------------
IN‐TEXT QUESTIONS
List four research instruments and explain two types of questionnaire.
Differentiate between focus group and tasting panel
Enumerate four limitations of interviews as research tool.
4.0 Conclusion
The most important consideration for collecting high-quality data and interpreting it accurately
knows what questions you want to answer with the information you gather. Once you have identified
those questions, you can choose the most appropriate methods for gathering and interpreting
information to answer them. With a focused effort, you will likely find the information you need to
create a successful research efforts.
5.0 SUMMARY
This unit covered areas of research tools / instruments for gathering data. We have
looked at various data collection tools such as anecdotal records, customer vendor feedback,
interviews (structured and unstructured), focus group, tasting panels, document analysis and
questionnaires. All these instruments are available for you to as Business education students in
your research endeavor and business field.
6.0 TUTOR MARKED ASSIGNMENT
Explain anecdotal records in relation to data collection in research.
Enumerate the basic guidelines for constructing a good questionnaire.
7.0 REFERENCES / FURTHER READINGS
Bennet, J. (2005). Evaluation Methods in Research. London: Continuum
Nworgu, B.C. (1991). Educational Research: Basic Issues and Methodology. Ibadan:
Wisdom Publishers Limited.
Kirk, R.E. (1990). Statistical: An Introduction. Forth Worth: TX Holt, Rinehart and
Wiston.
MODULE 3: STATISTICS IN EDUCATIONAL RESEARCH
Unit 1 Introduction to Statistics
Unit 2 Methods of Representing Data and Measures of Central Tendency
Unit 3 Measures of Variability or Spread
Unit 4 Measures of Association/Correlation
Unit 5 Testing of Hypothesis
Unit 6 Writing Research Reports
1.0 INTRODUCTION
In the previous modules/units, you worked through the different methods of collecting
data in research. The question is; what do you do with this seemingly unmanageable bulk of
data?
This question will take us to 'Data Analysis', which we shall describe "as the process of
organizing and summarizing data in order to provide answers to the research questions or test
hypotheses stated in the study". This process, most of the times, involves the use of statistical
procedures to summarize and describe the characteristics of samples and populations of the
study.
In this unit, we shall first look at the meaning of statistics, the types of statistics and
organization of data.
2.0 OBJECTIVES
At the end of this unit, you should be able to:
(i) define the concept statistics;
(ii) explain the types of statistics;
(iii) organize a set of scores under (a) sequencing, (b) frequency distribution table, (c) bar
chart.
3.4.1 Sequencing
This involves arranging the data in order of magnitude - ascending or descending order.
See example below:
Example 1:
Given that the test scores of 10 students in statistics are:
8, 9, 2, 5, 7, 6, 4, 9, 8, 3.
This could be arranged in ascending order thus:
2, 3, 4, 5, 6, 7, 8, 8, 9, 9 or in descending order thus; 9, 9, 8, 8, 7, 6, 5, 4, 3, 2.
If the data consists of names, they can be arranged in alphabetical order. If they consists
of objects, events, animals, etc. they can be arranged according to kinds, species, groups etc.
3.4.2 Tables
A table can be regarded as a two-dimensional representative of statistical information or
data. Tables can be simple or complex as shown in the examples on the enrolment of pupils in
central school Umuihi from 2000 to 2007, and Distribution of Mathematics teachers in Okigwe
Zone in the year 2006.
Example 1
Pupils' Enrolment in Central School, Umuihi, 2000 - 2007.
S/N Year Boys Girls Total
Example 2
Distribution of Mathematics Teachers in Okigwe Education Zone
S/N Local Government No, of Teachers
1. Ehime Mbano 525
2. Ihitte / Uboma 425
3. Isiala Mbano 600
4. Obowo – Etiti 400
5. Onuimo 325
6. Okigwe 425
Total 2,700
Example 3:
The scores of some students in a Mathematics test are given below. Present the scores in a
frequency table.
10, 15, 18, 12, 14, 15,20, 15, 16, 11, 12, 14, 19,20, 17, 18, 15, 13, 11, 12, 19, 13, 10, 14, 17, 19,
16,15,15,15.
Example 4:
Present the scores below in a grouped frequency table.
55, 62 60, 50, 52, 58, 55, 60, 51, 55, 68, 55, 47, 39, 58, 42, 47, 42, 48, 55, 48, 46, 55, 51, 58, 65,
52, 35, 54, 55, 52, 56, 46, 65, 53, 34, 48, 50, 39, 59, 53, 52, 33, 48, 65, 60, 36, 68, 45, 62, 59, 60,
33,40,61,38.
In order to determine the interval or class size:
(i) Find the range. This is given by the highest score minus the lowest score. From the
scores, we have 60 - 33 = 27.
(ii) Determine the number of groups. It has to be between 10 and 20.
(iii) Divide the range by the number e.g. 27 ÷ 12 = 3 (approximate)
(iv) Draw a table and tally the scores according to groups.
1.0 INTRODUCTION
In last unit, you were exposed to the concept of statistics and organization of data. You also read
through the bar chart which is a graphical way of representing data. In this unit, you will
continue to be exposed to other ways of representing data. These include pie chart, histogram,
frequency polygon and ogive. We will also look at the measures of central tendency. As this is
not a complete course on statistics, we may not be so detailed in the presentations.
2.0 OBJECTIVES
By the end of this unit, you should be able to:
(i) construct a pie chart using given data;
(ii) construct and describe histogram;
(Iii) draw a composite table and construct a frequency polygon;
(iv) draw a composite table and construct an ogive;
(v) calculate the mean, median and mode of a given data.
Example 1:
Construct a pie chart to represent the data below:
The distribution by local government area of Basic Technology teachers in Okigwe Zone is a
follows:
EHIME = 60, IHITTE/UBOM = 50, ISIALA = 65, ONUIMO = 40, OBOWO = 35, OKIGWE =
30.
To construct the pie chart:
i. Find the angle that is subtended at the centre by each group:
60 360
(a) EHIME = = 77.140
280 1
50 360
(b) IHITTE/UBOMA = = 64.290
280 1
65 360
(c) ISIALA = = 83.570
280 1
40 360
(d) ONUIMO = = 51.430
280 1
35 360
(e) OBOWO = = 45.000
280 1
30 360
(f) OKIGWE = = 38.570
280 1
ii. With the aid of a pair of compasses, any convenient radius, draw a circle.
iii. Using your protractor, mark out the angles corresponding to each group or category o
items, to the nearest degree.
i. Label the sectors of the circle corresponding to the items.
3.2 Histogram
In the last unit, you studied the bar chart, which is used mainly for the representation of
discrete data. In the construction, you noticed that the rectangles do not touch each other. The
histogram is used to represent data on a frequency distribution table like the bar chart. It is made
up of rectangular bars of equal joined to one another, and it is used for continuous data. At the
vertical axis, we have the frequencies and at the horizontal, we have the corresponding class
intervals. The difference between the two is that, for bar chart the class intervals are used while
for histogram the exact class boundaries arc used. There are two exact class boundaries—upper
and lower exact class boundaries. These are obtained by subtracting 0.5 from the upper boundary
and adding 0.5 to the lower boundary. Alternatively, for the exact lower limit of the first group
19 20
(20 - 24), we have = 19.5
2
24 25
And for the exact upper limit = = 24.5
2
Example 2
Using the data below, construct a histogram:
Class Interval 20-24 25-29 30-34 35-39 40-44 45-49 50-54 55-59 60-64 65-69 70- 74 75 -
79
Frequency 3 5 8 10 13 15 12 8 6 5 3 2
To construct a histogram:
1. Compose a composite table having the class interval, the exact class limits, and the
frequencies.
2. Choose suitable scales and draw the vertical and horizontal axes.
3. Mark of the frequencies on the vertical axis and the exact limits or real limits on the
horizontal axis.
4. Draw the rectangular bars on each boundary with the height corresponding to the
frequencies.
5. Draw arrows to show what is on the vertical and horizontal axes.
Registration = N5,000.00
Course materials = N10,000.00
Examinations = N5,000.00
Transportation = N3,000.00
Stationeries = N2,000.00
Diskettes and CDs = N1, 000.00
Note books = N2,500.00
Typing of assignments = Nl,500.00
Represent these expenses in a pie chart.
3.3 Frequency Polygon
This is a line graph plotted using the frequencies against the mid-points of the class
intervals.
Example 2
Use the data below to construct frequency polygon.
Class Interval 33-35 30-32 27-29 24-26 21-23 18-20 15-17 12-14 9-11
Frequency 3 5 8 10 13 15 12 8 6
Example 4:
Using the score groups below, draw an ogive or cumulative frequency curve:
Class Interval 5-9 10-14 15-19 20-24 25-29 30-34 35-39 40-44 45-49 50-54
Frequency 1 3 6 10 12 8 6 5 8 6
Example 5:
The scores often students in a test are as follows: 40, 55, 60, 30, 50, 48, 70, 85, 72, 65. Find the
mean.
To find the mean, we will add all the scores = X and divide by 10 = N i.e. 40 +
55+60+30+50+48+70+85+72+65.
X = 375. So X = X/N = 575/10 = 57.5
The mean can also be calculated from frequency distribution. In this case, we use the
formula: FX/F, where FX is the sum of the products of f and x and £F is the sum of
the frequencies.
Example 6:
Find the mean of the scores below:
S/N X F FX
1. 30 2 60
2. 20 4 80
3. 15 4 60
4. 25 3 75
5. 10 8 85
6. 8 2 16
7. 5 6 30
8. 21 2 42
9. 12 1 12
10. 24 5 120
37 575
Example 7:
Use the data given below to calculate the mean:
Class Interval 25-29 30-34 35-39 40-44 45-49 50-54 55-59 60-64
Frequency 2 5 6 7 10 6 3 2
(i) Complete the table by getting the mid-points X, and FX;
(ii) Use the same formula X = FX/F.
S/N Class Interval Mid-point (X) F FX
1. 60-64 62 2 124
2. 55-59 57 3 171
3. 50-54 52 6 312
4. 45-49 47 10 470
5. 40-44 42 7 294
6. 35-39 37 6 222
7. 30-34 32 5 160
8. 25-29 27 2 54
47 1807
X = 1807/41 = 44.07
You have seen that the mean can be calculated from both grouped and ungrouped
data, using different methods. One of these methods is called the assumed mean method.
It is called the short-cut.
Example 8:
Find the mean using the data on e.g. 7.
S/N Class Interval Mid-point (X) F X1 FX1
1 60-64 62 2 4 8
2 55-59 57 3 3 9
3 50-54 52 6 2 12
4 45-49 47 10 I 10
5 40-44 42 7 0 0
6 35-39 37 6 -1 -6
7 30-34 32 5 -2 -10
8 25-29 27 2 -3 -6
47 17
(i) Take away group mark as the assumed mean, and code it 0 as shown in column
X1.
(ii) Code every other mark above from 1, 2, 3 etc. and below -1, -2, etc.
(iii) Find the FX1 and sum up.
(iv) Use the formula AM + (FX1/F)i = X .
X = 42 + (17/41) = 42 + + 2.073 = 44.073
= 44.07
5 6 11
The median is = 5.5
2 2
When grouped data are given, the median is calculated using the formula
..
X L
N 2 cfb
i
fw
where L is the lower boundary of the median class;
N is the number of scores;
cfb is the cumulative frequency below the median class;
fw is the frequency within the median class.
Example 10:
Use the data below to find the median:
S/N Class Interval F FX
1. 85-89 1 52
2. 80-84 2 51
3. 75-79 3 49
4. 70-74 5 46
5. 65-69 7 41
6. 60-64 8 34
7. 55-59 10 26
8. 50-54 6 16
9. 45-49 5 10
10. 40-44 4 5
11. 35-39 0 1
12. 30-34 1 1
52
(iii)
..
X L
N 2 cfb
i
54.5
26 10
5
fw 10
= 54.5
105 = 54.5 + 5
10
= 59. 5
3.5.3 The Mode
This is the most frequently occurring score or scores in a distribution, It is the
most frequent score which can be easily determined by inspection. But in some
distributions, you may have two modes. This is called bimodal; any distribution with
more than two modes is called multi-modal.
Now, let us look at how to find the modes in the examples below:
Example 11:
Find the mode in the distribution below:
20, 30, 21, 45, 30, 25, 33, 35, 30, 22, 29, 30.
By inspection, you will see that 30 appeared 4 times. It is the mode because no other
score appeared up to that.
Example 12:
Find the mode in the frequency table given below:
X 10 9 8 7 6 5 4 3 2 1
F 1 2 4 5 8 6 4 3 1 1
Again, by inspection, you will see that the highest occurring frequency in the
above distribution is 8, and the value is 6. Therefore, 6 is the mode.
For a grouped data, the mode is calculated using the formula below:
..
X L
d 1 i
d1 d 2
where L is the exact lower limit of the modal class;
d1 is frequency of the modal class minus frequency of the class preceding or
before the modal class;
d2 is frequency of the modal class minus frequency of the class immediately after
the modal class.
Example 13:
Find the mode in the frequency table given below:
S/N Class Interval F
1. 85-89 3
2. 80-84 3
3. 75-79 8
4. 70-74 10
5. 65-69 12
6. 60-64 7
7. 55-59 5
8. 50-54 2
(iii)
..
X 64.5
55 64.5
52
52 7
= 64.5 + 3.571 = 68.07
IN-TEXT QUESTIONS
i. Define mean, median and mode.
ii. Find the mean, median and mode of the distribution given below:
10, 7, 8, 9, 6, 9, 3, 2, 9, 5, 1.
4.0 CONCLUSION
You have noticed that data by themselves convey little or no meaning until they
are summarised and described. Some methods of representing data have been presented
and the measures of central tendency, which form the central reference value that is
usually close to the point of greatest concentration of the measurement, and which may in
some sense be thought of typify the whole set, have also been presented. In the next unit,
we shall look at other statistical measures.
5.0 SUMMARY
In this unit, you have been able to go through the other methods of representing
data which you started in unit sixteen of this module. You have seen that the pie chart
uses a circle to represent a set of data or groups of items. In other words, it can be used
for both discrete and continuous data. You also went through the histogram, which is
made up of rectangular bars of equal width joined to one another, It is used for
continuous data. The frequency polygon is a line graph plotted using the frequencies
against the mid-points of the class intervals. The ogive uses the cumulative frequencies
against the exact "class boundaries. We have two types of ogives - 'less than' ogive and
'greater than' ogive.
You have equally worked through the measures of central tendency. The three
measures are the mean, the median and the mode. You have seen how to calculate these
measures. In the next unit, we shall look at other measures.
REFERENCES
Anaekwe, M. C. (2002), Basic Research Methods and Statistics in Education and Social
Sciences Enugu: Podiks Printing and Publishing Company.
Denga, L. D. and Ali A. (1983), An Introduction to Research methods and Statistics in
Education and Social Sciences. Jos: Savannah Publishers Limited.
Ogomaka, P. M. C. (1998) Descriptive Statistics for Research Students. Owerri:
Peacewise
Olatian, S. O. and Nwoke G. I. (1988) Practical Research Methods in Education. Onitsha:
Summer Educational Publishers.
UNIT 3: MEASURE OF VARIABILITY OR SPREAD
1.0 Introduction
2.0 Objectives
3.0 Main Content
3.1 The Range
3.2 The Quartiles
3.2.1 Calculation of the Quartiles
3.2.2 Interquartile Range
3.2.3 Quartile Deviation or Semi-interquartile Range
3.3 The Percentiles
3.3.1 The Deciles
3.4 The Variance and the Standard Deviation
4.0 Conclusion
5.0 Summary
6.0 Tutor Marked Assignment
7.0 References and Further Readings
1.0 INTRODUCTION
In the last unit, you worked through the measures of central tendency. In addition
to those measures, researchers are also interested to know how the scores are spread or
scattered in the distribution.
So the measures of variability indicate the degree to which a set of scores differs
from each other in the distribution. These measures present a measure of homogeneity
within the group of scores.
In this unit, we shall look at the range, the quartiles, the percentiles, the variance
and the standard deviation.
2.0 OBJECTIVES
After working through this unit, you will be able to:
(i) find the range in a given set of scores;
(ii) explain and find the quartiles in a distribution;
(iii) find the percentiles in a given set of scores;
(iv) calculate the variance in a given set of scores;
(v) calculate the standard deviation in a distribution.
Example 1:
Find the range of the scores below:
30, 45, 20, 32, 70, 85, 90, 44, 60.
You will notice that the lowest score is 20 and the highest score is 90. So, Xh – XL + 1 =
90 – 20 + 1 = 71. The range is 71.
You would have seen that the range is affected by the two extreme scores.
Therefore, it is an unstable and unreliable method of determining the spread of scores.
Because of this limitation, it is seldomly used as an indicator of the spread.
3.2 The Quartiles
These are score points or values which subdivide a given distribution into four
equal parts. In other words, the number of scores in anyone of the four groups is equal to
the number of scores in any other of the remaining three groups.
There are only three quartiles for any given 'distribution. These are the first
quartile Q1, second quartile Q2 and third quartile Q3. This can be illustrated below;
Qi L
iN 4 cfbc
fw
Where i = 1, 2, 3, (i.e. the quartiles)
N = f = sample size
L = lower class boundary of the quartile class
cfb = cummulative frequency below the quartile class
fw = frequency of the quartile class
c = class interval size.
Example:
Find Q1 and Q3 in the distribution below:
S/N Class F
Interval
1. 50 – 54 1 34
2. 45 – 49 2 33
3. 40 – 44 2 31
4. 35 – 39 5 29
5. 30 – 34 8 24
6. 25 – 29 6 16
7. 20 – 24 4 10
8. 15 – 19 3 6
9. 10 – 14 2 3
10. 5–9 1 1
34
(X X )2 (X X )2
frequencies i.e. f or f . This is the deviation method.
f f
There is also the raw score method otherwise called the machine approach. We shall
look at it after the deviation method. Now, let us take some examples.
S/No X F fX (X X ) (X X )2 f (X X )2
1 11 1 11 -4.97 24.70 24.70
2 12 2 24 -3.97 15.76 31.52
3 13 4 52 -2.97 8.82 35.28
4 14 7 98 -1.97 3.88 27.16
5 15 10 150 -0.97 0.94 9.40
6 16 12 192 0.03 0.00 0.00
7 17 11 187 1.03 1.06 11.66
8 18 6 108 2.03 4.12 24.72
9 19 4 76 3.03 9.18 36.72
10 20 3 60 4.03 16.24 48.72
60 958 249.88
Example
Find the variance and standard deviation of the following scores:
fx 958
Steps: (i) Find the mean = = = 15.97
f 60
Step:
(i) Complete the composite table as shown.
13 13 17 7 22 22 26 17 13
14 16 7 6 18 20 10 17 11
10 17 11 10 15 16 8 16 21
7.0 REFERENCES
Anaekwe, M. C.(2002) Basic Research Methods and Statistics in Education and Social
Science. Enugu: Podiks Printing and Publishing Company.
Denga, I. D. and Ali, A. (1983): An Introduction to Research Methods and Statistics in
Education and Social Sciences, Jos: Savannah Publishers Limited.
Ogomaka, P.M.C (1998), Descriptive Statistics for Research Students. Owerri:
Peacewise.
Olatian, S. O and Nwoke, G. I. (1988). Practical Research Methods in Education.
Onitsha: Summer Education Publishers.
UNIT 4
MEASURES OF ASSOCIATION/CORRELATION
1.0 Introduction
2.0 Objectives
3.0 Main Content
3.1 The Concept of Correlation
3.2 Scatter-grams of Various Correlations
3.3 Pearson Product Moment Correlation Coefficient (r)
3.3.1 Calculating Pearson r using Deviations from the Mean
3.3.2 Calculating Pearson r using the Raw Score Method
3.4 Spearman Rank Order Correlation Coefficient - rho
3.4.1 Calculation of Spearman Rank Order Correlation
3.5 Point Biserial Correlation Coefficient - rpbi
4.0 Conclusion
5.0 Summary
6.0 Tutor Marked Assignment
7.0 References and Further Readings
1.0 INTRODUCTION
In the previous units, we have focused on sample scores from one variable or
distribution of scores from one variable.
In this unit, you will learn about matched or paired sets of scores. Specifically,
measures of association show the degree of relationship between two or more variables.
We shall be looking at some of these measures or the statistics for describing the
extent of correlation or 'going together' of some attributes or characteristics possessed by
a sample of individuals. This degree of relationship between the attributes or variables is
expressed as a coefficient of correlation.
The result of this unit will teach you the most common types of correlation which
are Pearson .Product Moment and Spearman Rank Order.
2.0 OBJECTIVES
At the end of this unit, you will be able to:
(i) define correlation;
(ii) illustrate the scatter-grams of various correlations;
(iii) calculate the Pearson r;
(iv) calculate the Spearman rho.
(e) r = No Relationship
3.3 Pearson Product Moment Correlation Coefficient (r)
This type of correlation coefficient, named after the man who developed it, is
used when the two sets of data are continuous or interval data. There are two major
approaches of calculating the Pearson Product Moment correlation coefficient (r).
The first is the deviations from the mean approach, while the second is the raw scores
approach. Let us look at them one after the other.
where x = X- X , y = Y- Y
Example 1:
Using the data below, calculate the Pearson r.
X 10 11 12 12 13 14 15 15 16 17 17 18 18
Y 5 8 9 4 7 6 8 9 10 10 12 14 13
Step:
(i) Find the mean for X and Y.
(ii) Complete the composite table.
X Y X-X Y-Y XY X2 Y2
(x) (y)
1 10 5 ^.5 -3.8 17.10 20.25 14.44
2 11 8 -3.5 -0.8 2.80 12.25 0.64
3 12 9 -2.5 0.2 -0.50 6.25 0.04
4 12 4 -2.5 -^.8 12.00 6.25 23,04
5 13 7 -1.5 -1.8 2.70 2.25 3.24
6 14 6 -0.5 -2.8 1.40 0.25 7.84
7 15 8 0.5 -0.8 -0.40 0.25 0.64
8 15 9 0.5 0.2 0.10 0.25 0.04
9 16 10 1.5 1.2 1.80 2.25 1.44
10 17 10 2/5 1.2 3.00 6,25 1.44
11 17 12 2.5 3.2 8.00 6.25 10.24
12 18 14 3.5 5.2 18.20 12.25 27.04
13 18 13 35 4.2 14.70 12.25 17.64
188 115 80.90 87.25 107.72
14.5 8.8
r = 0.83
3.3.2 Calculating Pearson r using the Raw Score Method
N XY X Y
The formula is given by r =
N X 2 X 2 N Y 2 Y
2
where x = X- X , y = Y- Y
Example 2:
Let us use the same data in example 19.1.
X 10 11 12 12 13 14 15 15 16 17 17 18 18
Y 5 8 9 4 7 6 8 9 10 10 12 14 13
Steps:
i. Complete the composite table.
ii. If N = 13, X = 188, Y = 115, XY = 1744, X2 = 2806 and Y2 = 1125, then:
N XY X Y
r =
N X 2 X 2 N Y 2 Y
2
S/N X Y XY X2 Y2
1 10 5 50 100 25
2 11 8 88 121 64
3 12 9 108 144 81
4 12 4 48 144 16
5 13 7 91 169 49
6 14 6 84 196 36
7 15 8 120 225 64
8 15 9 135 225 81
9 16 10 160 256 100
10 17 10 170 289 100
11 17 12 204 289 144
12 18 14 252 324 196
13 18 13 234 324 169
188 115 1744 2806 1125
14.5 8.8
13 1744 188 115
r =
13 2806 188 2 13 1125 115 2
22672 21620
=
36478 35344 14625 13225
1052
=
1134 1400
1052
= = 0.83
1260
You can see that the two approaches give the same result. This is because the
formula of the raw scores method is derivable from the formula of the deviations from
the mean method. You will have to note that when the scores are large and the means of
X and Y are whole numbers, the deviations from the mean method becomes simper to
handle. But when the means of X and Y are not whole numbers the raw score method is
preferred.
IN-TEXT QUESTION
Use any method to calculate the Pearson r of the data:
S/N 1 2 3 4 5 6 7 8 9 10
X 51 44 70 32 65 67 19 71 45 80
Y 49 41 45 31 50 61 11 64 21 75
Example 3:
Calculate the rho of the data presented below:
S/N 1 2 3 4 5 6 7 8 9 10
X 51 44 70 32 65 67 19 71 45 80
Y 49 41 45 31 50 61 11 64 21 15
S/N X 'Y RX RY D D2
1 51 49 6 5 1 1
2 44 41 8 7 1 1
3 70 45 3 6 -3 9
4 32 31 9 8 1 1
5 65 50 5 4 1 1
6 67 61 4 3 1 1
7 19 11 10 10 0 0
8 71 64 2 2 0 0
9 45 21 7 9 -2 4
10 80 75 1 1 0 0
10
Steps:
i. Complete the composite table by getting the ranks and the differences between the
ranks.
6 D 2
ii. Apply the formula: rho = 1
N N2 1
6 18 108 108
rho = 1 = 1 = 1
10 10 1
2
10 99 990
= 1 – 0.109 = 0.891
Xp Xq Xp Xt
The formula for this is given by: rpbi = pq p q
St St
where Xp = mean score of the continuous variable of the subgroup
that belongs to the natural dichotomy p.
Xq = mean score of the continuous variable of the subgroup
that belongs to the natural dichotomy q.
st = standard deviation of the total scores for the whole group
on the continuous variable.
p = proportion of the number of members in subgroup
p.
q = proportion of the number of members in subgroup q.
Now, let us look at the steps you can follow:
i. Find Xp = mean for the proportion of boys and the group.
15 13 20 16 17 09 07 97
= = 13.86
7 7
ii. Find Xq = mean for the proportion of girls in the group.
10 11 12 18 14 65
= = 13.0
5 5
iii. Find p= 7/12 = 0.58
iv. Find q = 5/12 = 0.42.
v. Find St.
S/N X X-X (X - X )2
1 10 -3.5 12.25
2 15 1.5 2:25
3 11 -2.5 6.25
4 13 -0.5 0.25
5 12 -1.5 2.25
6 18 4.5 20.25
7 20 6.5 42.25
8 14 0.5 0.25
9 16 2.5 6.25
10 17 3.5 12.25
11 09 -4.5 20.25
12 07 -6.5 42.25
x 162 13.5 167.00
X X
2
167
St = r = = 13.92
n 12
= 3.73
Xp Xq 13.86 - 13.0q
rpbi = pq = 0.58 0.42
St 3.73
= 0.2305563 x 0.493558
= 0.1137963 = 0.11
IN-TEXT QUESTION
Find the rpbi of the following data:
S/N i 2 3 4 5 6 7 8 9 10 11 12 13
X 60 40 55 20 70 35 48 15 30 57 65 25 30
Y G B B G B G B G B G B G G
4.0 CONCLUSION
A very good number of research studies tend to determine the nature and scope of
relationships which exist between two or more variables being investigated. In this unit,
you have seen that the degree of relationship which exists between variables is referred to
as correlation. You have also noted that the statistical index of measuring the relationship
is called correlation coefficient. This correlation coefficient presents a picture of how a
change in one variable results in a change in the corresponding correlated variable. The
result of the correlation tests can be used for predictive purposes. But they cannot be used
for establishing a cause-effect relationship between two variables.
5.0 SUMMARY
In this unit, you have learnt that correlation is the extent or degree of relationship
between two variables while the index showing the degree of such relationship between
the two variables is i called correlation coefficient Correlation values range from-1 to+1.
Scatter-grams of different [types of relationships were shown. Pearson Product Moment
Correlation Coefficient otherwise called Pearson r was also discussed with the two
methods for the computation. These are the deviation and the raw score methods. The
methods for calculating the Spearman rho and the Point Biserial Correlation (rpbi) were
discussed in detail. The next unit will take us to the test of hypotheses to complete the
module.
1.0 INTRODUCTION
In Module one, you were introduced to the types of hypotheses. In this unit, you will
learn how to test the hypotheses using some of the statistical tests.
The purpose of testing a hypothesis is to determine the probability that it is supported by
facts. You may test a hypothesis by applying it to already known facts or taking it as a new
appeal to experience. This same mental technique to problem-solving is also employed by
science and philosophy.
Hypotheses are used as indicators of the realistic answers which researchers have to their
stated problems or questions in research. So when hypotheses are tested, the results lead to
establishment of new facts or confirmation of old ones. If a hypothesis is successfully verified or
tested and confirmed to be true, it is then used to support a theory.
In other words, theories are developed, tested and confirmed in research through the
process of hypothesis testing. This leads to the generation or advancement of knowledge. In this
unit, you are going to be exposed to the rudiments of the processes involved in testing
hypotheses,
2.0 OBJECTIVES
After going through this unit, you will be able to:
(i) explain the alpha level or level of confidence and degree of freedom;
(ii) discuss the two types of errors in hypothesis testing;
(iii) use the t-test to test a given null hypothesis;
(iv) use the relationship between correlation coefficient and t-test in hypothesis testing;
(v) use analysis of variance to test hypothesis;
(vi) use chi-square to test hypothesis;
(vii) explain the meaning of one-tailed and two-tailed tests.
3.0 MAIN CONTENT
3.1 Selection of the Level of Significance or Alpha Level
In proposing your hypothesis, you must include a confidence limit, otherwise called alpha
level () or significance level.
In most researches in education, two alpha levels are used. These are 5% (0.05) and 1%
(0.01). If you choose 5% in a particular study, the implication is that if your study is replicated
100 times, the same outcome will occur 95 out of 100, and 5 out of 100 may vary due to chance.
If it is 1% (0.01) level of significance, it means that if your study is replicated 100 times, you are
sure 99 out of 100 will be correct while 1 out of 100 may vary due to chance factors. This is a
more rigorous confidence level.
At this point, you need to note that when you test a hypothesis, you are only dealing with
probability of something being true or false. Hypothesis testing helps you to make predictions '
and not to establish causation. It does not provide absolute proof. In other words, a hypothesis
cannot be proved absolutely true or false.
At this point, you have to take a decision. This will be based on the comparison of the
calculated value oft-test and the value oft-test on the table or the critical region.
Now that tcal = 2.795, df - 25 - 1 = 24, alpha level - 0.05
ttabat(25 : 0.05) = 2.060.
For decision rule, if calculated value is greater than the value in the table or critical value,
Reject the null hypothesis. But, if the calculated value is less than the value on the table, Accept
ho.
From this result, tcal is greater than the ttab i.e. 2.795 > 2.060. We therefore reject that
there is no significant difference between the population mean and the sample, mean. It implies
that there is a significant difference between the two means.
IN-TEXT QUESTIONS
What do you understand by the following:
(a) P < 0.05
(b) degree of freedom
(c) Type I and Type II errors
(d) Two tailed and one tailed test.
n 1 n 2 2n 1 n 2
where S1 =
X 1 X2
n1 1
Example 2:
A teacher wanted to compare the academic performance of two sets of students in his school
with a view to finding out whether their mean performances are significantly different. He
collected samples of the two sets. His results are shown in the table below:
Set Mean Performance Standard deviation No. of Samples
2005 50% 14.50 80
2006 58% 12.00 75
Solution:
i. Propose a null hypothesis H0: There is no significant difference between the mean
performances of the students from the two sets.
XX
ii. t =
n 1 1S1 n 2 1S 2 2 n 1 n 2
2
n 1 n 2 2n 1 n 2
58 50
t =
75 112 2
80 114.5 2 75 80
75 80 275 80
8
=
74144 79210.25155
153 6000
8
=
10656 16609.75155
918000
8
=
27265.75 155
918000
8
=
4226191.3
918000
8 8
= =
4.6036942 2.1456221
= 3.7285224 = 3.73
iii. Decision:
tcal = 3.73, ttab at (75 + 80 – 2 : 0.05/2) = ttab at 153 : 0.05
tcal = 3.73. t(153:0.025) = 1.96
Since tcal is greater than ttab, we reject H0. It means that there is a significant difference
between the mean performances of the two sets of students.
IN-TEXT QUESTIONS
The result of a researcher’s study, to find out if there is a significant difference between the
performances of males and females in his class is given below:
Gender Mean Performance Standard deviation No. of Samples
Males 65% 11.50 45
Females 58% 14.20 40
Are the gender-wise performances significantly different?
where d
d , S = standard deviation of the ds.
n
Example 3:
A set of students took tests in both Mathematics and Statistics. Their results are as follows:
S/N 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
Mathematics 50 65 70 35 44 52 67 72 48 38 59 65 62 40 54 64 70 55
Statistics 48 60 74 30 40 50- 69 70 50 42 60 70 60 29 52 61 70 53
i. Complete the table by getting d = difference (linear) between Mathematics and Statistics.
S/N Mathematics Statistics D d2
1 50 48 2 4
2 65 60 5 25
3 70 74 -4 16
4 35 30 5 25
5 44 40 4 16
6 52 50 2 4
7 67 69 -2 4
8 72 70 2 4
9 48 50 -2 4
10 38 42 -4 16
11 59 60 -1 1
12 65 70 -5 25
13 62 60 2 4
14 40 29 11 121
15 54 52 2 4
16 64 61 3 9
17 70 70 0 0
18 55 53 2 4
22 286
Example 4:
A teacher wanted to find out whether students' scores in Technical Drawing have any significant
relationship with their scores in Mathematics. He used the Pearson Product Moment Correlation
Coefficient to do this. He came out with a correlation coefficient of r - 0.60, N = 50.
To find out if this is significant:
i. Propose a null hypothesis: The students' scores in Technical Drawing and Mathematics
are not significantly related. OR, There is no significant relationship between the
students' scores in both Mathematics and Technical Drawing.
ii. Substituting with the formula:
n2 50 2 0.60 48
t = = 0.60 =
1 r 2
1 0.60 2 0.64
4.1569219
=
0 .8
= 5.196
iii. Find the critical value by using t(so- i: o.Q5) = 2.021.
iv. Since tcal is greater than ttab i-e. 5.196 > 2.021, we reject the null hypothesis and say the
students* scores in Mathematics and Technical Drawing are significantly related.
IN-TEXT QUESTION
In a research study, it was found that the correlation coefficient of two variables was
0.72 and the number of the respondents was 50. Propose a null hypothesis and test it
using this information at 0.05 levels.
Example 5:
Scores of three randomly selected groups of students in an English test are given below.
GP 1 15 20 12 10 9 7 6 11 18 14 5
GP2 13 12 15 19 20 11 8 14 10 9 4
GP3 18 16 13 9 8 4 20 18 12 7 10
Test the hypothesis that the three groups do not belong to the same population.
S/N X1 X2 X3 X12 X22 X32
1 15 13 18 225 169 324
2 20 12 16 400 144 256
3 12 15 13 144 225 169
4 10 19 9 100 361 81
5 9 20 8 81 400 64
6 7 11 4 49 121 16
7 6 8 20 36 64 400
8 11 14 18 121 196 324
9 18 10 12 324 100 144
10 14 9 7 196 81 49
11 5 4 10 25 16 100
127 135 135 1701 1877 1927
X 11.55 12.27 12.27
Find:
1. X1= X1 + X2 + X3 = 127 + 135 + 135 = 397
2 2 3 3
2. X 1= X 1 + X 2 + X 3 = 1701 + 1877 + 1927 = 5505
3. N1 = N1 + N2 + N3 = 11 + 11 + 11 = 33
X 1
2
397 2
We shall take the correct factor to be =
N1 33
X 2
157609
X
1
= 5505
2
4. Sum of squares total (SSt) = 1
N1 33
= 5505 – 4776.03 = 728.97
5. Sum of squares, between group (SSb):
X X X X
1
2
1
2
1
2
1
2
SSb =
N1 N2 N3 N1
127 2 135 2 135 2 397 2
=
11 11 11 33
= 1466.2727 + 1656.8182 + 1656.8182 – 4776.03
= 4779l9091 – 4776.03 = 3.891
6. Sum of squares within groub (SSw): SSw = SSt – SSb = 728.97 – 3.8791
= 725.09
7. Degree of freedom, between (dfb) = K – 1 = 3 – 1 = 2.
8. Degree of freedom, whiting (dfw) = N – K = 33 – 3 = 30.
Where N = total number of sample
SS b 3.8791
9. Variance, between groups (Vw) = =
df b 2
= 1.94
SS w 725.09
10. Variance, within groups (Vb) = =
df w 30
= 24.17
Vb 1.94
11. F – ratio = = = 0.08
Vw 24.17
12. Determine the critical value of F.
From the calculation dfb, = 2 and dfw = 30, go to the F-table and find the point of
intersection of 2 and 30 at 0.05 level. This will give you the F-value i.e. 3.32.
13. Decision: F - value calculated = 0.08
F - value critical = 3.32
Since the calculated value is less than the critical value for the degrees of freedom 2 and 30, and
alpha level of 0.05, we ACCEPT the null hypothesis that the scores are not significantly
different.
For the purpose of presenting the result in a research report, a summary of the results is
shown in a table while the computations are shown in the appendix. Thus:
Sources of Sum of Degree of Variance Fcal F-crit Decision
Variation squares freedom
Between groups 3.8791 2 1.94 0.08 3.32 Accept ho
Within groups 725.0900 30 24.17
Total 728.9691 32
IN-TEXT QUESTION
S/N 1 2 3 4 5 6 7 8 9 10
X1 6 7 13 8 12 5 10 6 9 11
X2 15 14 10 12 13 11 14 10 12 13
X3 5 8 10 15 4 13 7 13 6 9
X4 10 7 5 8 9 8 6 4 7 3
Example 6:
A survey to determine the preference pattern of some parents on the choice of courses for their
children is given in a table below. Use a null hypothesis to determine whether the indicated
preference pattern is statistically significant.
Frequency Business Law Medicine Engineering Total
Observed 24 50 52 34 160
Expected 40 40 40 40 160
Steps:
i. State the null hypothesis H0: There is no significant difference between the expected and
observed preference pattern of the parents at 0.05 alpha levels.
ii. Apply the chi-square formula in each cell and sum up at the end.
1. For Business =
O E 2 =
24 40 2 = 6.4
E 40
2. For Law =
O E
2
=
50 40
2
= 2.5
E 40
3. For Medicine =
O E 2 =
52 402 = 3.6
E 40
4. For Engineering =
O E
2
=
34 40
2
= 0.9
E 40
O - E
2
x 2
= 6.4 + 2.5 + 3.6 + 0.9 = 13.4
E
To take decision on the significance of the x2 value, you have to find the degree of
freedom df. The example discussed above is a one-variable case, so the df is given by: df = K - 1,
i.e. (4 - 1) = 3. As usual, go to the chi-square table and look under df = 3, and your alpha level,
which can be 0.05 or 0.01. Again, if the calculated value exceeds the value on the table, you
reject the null hypothesis. In this case x2 at 3 : 0.05 - 7,82. This is less than the calculated value,
so we reject the null hypothesis.
Most of the times, researchers are confronted with the test for the independence of two
variables. For instance, gender and opinion, or religion and choice or age and opinion. Again,
each of the
variables may have two or more levels. The observed and the expected frequencies arc presented
in a table called contingency table. It has a number of rows and columns.
Example 7:
The enrolment pattern of students to different academic programmes according to religion is
given in the table below. Calculate the chi-square and test for the significance at 0.05.
Religion Academic Programmes
Business Law Medicine Engineering Totals
Christianity 50 35 48 45 178
Islam 30 45 35 50 160
Traditional 45 30 25 40 140
Godian 25 20 30 28 103
Totals 150 130 138 163 581
To solve this problem, take the table above as the table of the observed frequencies. Therefore,
you will need the table for the expected frequencies. To find the expected frequency for each
cell, apply the formula:
column total row total
overall total
150 178
Example, for cell 1, where the observed is 50, the expected is given by = 45.96.
581
130 178
For the next cell where the observed as 35, the expected is given by = 39.83 etc.
581
The expected frequencies are better presented in a table like the observed. See the table below
IN-TEXT QUESTION
Use the data below to verify your proposed null hypothesis:
Gender VX VY VZ Total
Male 55 40 50 145
Female 35 25 40 100
Total 90 65 90 245
4.0 CONCLUSION
Now that you have successfully worked through this unit on how to test hypotheses, you
are now prepared to carryout your research project work. But before you go properly into that,
we shall introduce you to how to write research reports in the next unit.
5.0 SUMMARY
In this unit, we have discussed the selection of the alpha level or significance level and
we said the two most common alpha levels used in research are 0.05 and 0.01. We touched upon
the degrees of freedom, Type J error and Type II error as the likely errors that can be made in
decision making in the test of hypothesis.
Hypotheses can be frame in two formats, which are directional and non-directional. This
implies that we have two types, vis-a-vis one tailed test and two tailed test. You also studied the
different types of tests used in testing hypotheses. The t-test, the F-test and the chi-square are the
prominent.
In the next unit, you will be introduced to how to write your research reports.
1.0 INTRODUCTION
The final stage of any research process is the writing of the research report. Research is
very important, because the findings generated can be used for rational decision-making and,
according to Nkpa (1979), as a springboard for further research. The main aim of writing
research report is to communicate or disseminate the research findings to the literate audience. In
writing the research report, the impersonal mode is preferred. That is to say, instead of say “I did
this”, you should say “the study was carried out to do this”.
You will have to note that in presenting a research report, you have to use the required
format. Most institutions have their own format. These formats or house-styles do not vary
significantly from the general format. National Open University of Nigeria (NOUN), School of
Education, has its own house-style. You supervisor will let you have it. For the purpose of this
unit, we shall discuss the general format.
2.0 OBJECTIVES
At the end of this unit, you should be able to:
(i) Itemise the chapter titles and sub-titles in a research project report;
(ii) Prepare a research report based on the given format.
3.2.2 Introduction
i. Background to the Problem: Here, such factors or circumstances that informed the
investigation are traced. It is presented using reasoned statements to show that it is
worthwhile to dissipate resources to carryout the study. It shows the nature, scope, and
current status of the factors affecting the problem.
It has to be presented in a way as to be clear and convincing to the reader.
ii. Statement of the Problem: The problem as already highlighted is stated in simple, clear
and unambiguous terms. This is not required to be too long.
iii. Purpose of the Study/Objectives of the Study: These go interchangeably, but it states the
specific aspects of the study and the reasons for focusing on them. It includes statements
of what should be accomplished and all that would be investigated.
iv. Significance of the Problem: The usefulness, the utility value of the research or findings
of the research should be articulated. The institutions, groups or individuals who are
expected to profit or benefit and the benefits expected to accrue to them are to be stated
in this section.
v. Scope of the Study: This is interchanged with the delimitation of the study. Here, you
will have to indicate the extent to which the study will be covered. It involves the
geographical area, time period, and variables to be covered.
vi. Research Questions and/or Hypotheses: These are formulated based on the type of
research and the variables under investigation. They should be formulated to provide
answers to the problems under study.
vii. Definitions of Terms: The essence of definition is to ensure that the reader understands
the specific meanings ascribed to the terms by the author So you have to use this to
educate the readers on the operational meaning of any coined, technical words, phrases or
expressions which cannot otherwise be understood because of their unconventional
usage.
4.0 CONCLUSION
At the end of your programme, you are expected to carryout a research. At the end of the
research, you are also expected to submit a written report of the investigation. In this unit, you
have gone through the involvement in the writing of the report. A very important demand here is
that you must be as objective as possible in your report. At the initial stage, you cannot make any
statement that would show you are in favour or against an idea. Your report should be devoid of
emotional or subjective statements. You should arrange the different parts of the report so as to
make it possible for a reader to easily locate any section of particular interest to him,
5.0 SUMMARY
In this unit, we have discussed and presented a sample format of a research report. We
have also discussed these steps in details stating from the preliminary stages to the
supplementary stages. We have emphasized that your reports should not be presented with
personal pronouns like I, my, we etc. Instead use impersonal pronouns and passive voice. You
should make sure that the report is written in a clear, simple and straightforward style. Your
motive should be effective communication. Therefore, use very simple language. You should
always be brief so as not to bore your reader. Abbreviations should only be used after they have
been written in full earlier. Avoid the use of generalizations or conclusions, which are not
supported by the findings, We also said that every source cited in the work or used but noted
cited in the work should be documented in the reference page. Improper citation or inability to
give details of a source cited in the body of the work should be documented in the reference
page. Improper citation or inability to give details of a source cited in the body of the work
should be avoided. Remember that proofread the report thoroughly after typesetting. This will
help you not submit avoidable errors.
Congratulations for being part of the success story of NOUN, and for graduating in this
programme.