An Overview of Content Analysis
An Overview of Content Analysis
com
Introduction
1
Correspondence: Tracy G. Harwood, Leicester Business School, De Montfort
University, The Gateway, Leicester, LE1 9BH, Tel: 0116 257 7214, Fax: 0116 257
7204, Contact: [email protected], http://www.dmu.ac.uk/business
Basic Concepts
IP: 119.92.61.209 On: Sat, 10 Oct 2015 22:04:28
the scheme to be used. That is, all the data classified into the same category
refers to the same object or event and everything which has been excluded
must not refer to that object or event. Various levels of classifying the
content of communications have evolved (Janis 1965) where the meanings of
words are inferred from the context in which they have occurred relative to
accompanying behaviours –
IP: 119.92.61.209 On: Sat, 10 Oct 2015 22:04:28
Techniques
Systems
Krippendorff (1980) defines a system as a conceptual device describing a
portion of reality. Broader descriptions are “an entity that maintains its
existence and functions as a whole through the interaction of its parts”
(O’Connor 1996, p 1); “a boundary-maintaining set of interdependent
482 Tracy G. Harwood and Tony Garry
Standards
Standards allow comparison of objects through relative measurement by
identifying, evaluating and auditing the content of communication. Evaluation
is somewhat subjective since many studies have lacked defensible criteria for
measurement against some standard (Krippendorff 1980). Audits are also
subject to judgment although the standard may be intertwined with legal or
professional directives, eg., the Advertising Standards Association, against
whose guidelines of ‘legal, decent, honest and truthful’ advertisements for
non-broadcast media are deemed fit for public consumption
(www.asa.org.uk/). Identification of a given characteristic is, however, easier
since it either does or does not exhibit that characteristic. Janis (1965)
argues most content analyses involve the categorisation of ‘sign vehicles’,
based on meanings interpreted by the receiver. This is at least partially due
An Overview of Content Analysis 483
Indices
Delivered by Publishing Technology to: Chinese University of Hong Kong
Linguistic representations
Discourse (the sequence of a communication) is of considerable interest
since reference to an utterance alone obviates the connotation and
understanding of language. That is, behaviours exhibited may contain a
series of lexical items with essentially similar meanings in isolation, however,
in their context may have entirely different meanings. An example from Hays
(1969) compares the content of three sentences: “the door opened”; “the
janitor opened the door”, “the key opened the door” (p 62). Here the subject
of each sentence is different (door, janitor, key) but the action of opening the
door is the same. It is the relationship between the subject and the event
which changes.
Linguistic representation is, therefore, a method of content analysis
whereby language is classified according to qualitative references to, rather
than a quantitative index of, an object. This has been used by researchers to
develop models of behaviour to analyse, for example, options remaining
during arms limitation negotiations (Allen 1963). Developments in computer-
aided content analysis also enhance the opportunity to analyse inferences,
objectives, disagreements, or alternative courses (Hays 1969; Krippendorff
1980; Wilson and Rayson 1996).
Communications
Communication takes place in the context of an existing relationship
among subjects and, in so doing, may alter that relationship. Although rarely
used in content analysis, this technique has been utilised to analyse
response and cue elements of interactions by Donohue et al. (1984) within
the context of small groups. A more common methodology for analysis of
communication, however, is interaction analysis (Bales 1950; Krippendorff
1980).
484 Tracy G. Harwood and Tony Garry
Institutional processes
Verbal communications within organisations have many different purposes
and meanings depending on structure and processes. Face value may,
however, be open to a number of interpretations, particularly when viewed by
an external party. Indeed, organisational communications are governed by
Delivered by Publishing Technology to: Chinese University of Hong Kong
and the validity of the categories or coding mechanism that is used in a given
context. This paper now reviews these areas in some detail. Figure 1
provides an overview of a typical content analytic research design, based on
Neuendorf (2002).
Reliability
Delivered by Publishing Technology to: Chinese University of Hong Kong
behaviour. This is not surprising given the apparent relative ease of unitising
data, which may be by one of the following means –
(Kolbe and Burnett 1991) while the process of training of judges may also
offer opportunities to pre-test identified categories through discussion and
consensus. It is also the case that unit sizes may not be appropriate for the
categories chosen. For example, units may be too small, resulting in an
increase in unclassifiable data; or too large, resulting in ambiguities between
judges.
Measures of reliability enhance the research findings, supporting
decisions, drawing inferences from or making recommendations based on
the original data. Such measures also enable the methodology to be used
quantitatively (Guetzkow 1950). Researchers have, however, been criticised
for their under-use of rigorous statistical calculation of category and judge
reliability (Kolbe and Burnett 1991), raising obvious questions on credibility.
some are rarely used. A measure developed to remove the impact of chance
agreement is Cohen’s (1960) kappa although it is difficult to calculate where
chance is unknown or cannot be approximated. Kappa may be tested for
significance (Bakeman and Gottman 1987) and is also a commonly used
measure in sequential analyses, where the sequence of categories is of
Delivered by Publishing Technology to: Chinese University of Hong Kong
Validity
Validity is the extent to which the research may be generalised to the
population. Two types of validity exist in relation to content analysis (Weber
1990) –
analyse interview transcripts (Wilson and Rayson 1996); QSR NUDIST for
‘non-numerical unstructured data indexing searching and theorizing’,
supporting processes such as managing, exploring and searching the text of
documents; manage and explore ideas about the data; link ideas and
construct theories about the data; test theories about the data; and, generate
Delivered by Publishing Technology to: Chinese University of Hong Kong
An example is SAS Text Minor for ‘statistical analysis system’, launched mid-
2002 (www.sas.com). Text Miner can analyse the content of large text-
based files such as emails, call reports and other documents and assist in
uncovering patterns in data. It is generally used in the context of customer
relationship management (CRM) programs wherein managers receive and
process large amounts of text-based information; human resources where it
is used for reviewing and matching applicant curriculum vitaes; and, analysis
of open-ended survey responses to develop hypotheses and identify trends.
It enables full text ‘pre-processing’ by integrating unstructured textual data
with existing structured data such as demographic and psychographic
purchasing data.
As intimated in the preceding text, content analysis can be applied at
differing levels of complexity and so, therefore, this article would not be
complete without consideration of the levels of complexity required to capture
phenomena of interest. Clearly, as a method for observation, it is possible to
design a mechanism which captures both verbal and non-verbal behaviour.
Whilst verbal behaviour is relatively easy to capture with recording equipment
eg., tape, video, pencil and paper, non-verbal behaviour provides
researchers with a more complex set of considerations, particularly if the
phenomena of interest are only to be observed in real-life. Consider, for
example, the resource required to capture in sufficient detail the physical
movements and expressions of every participant in a sales negotiation as
unobtrusively as possible. Of course, modern technology can provide
solutions to this dilemma but there then remains the problems of deciphering
the meanings of non-verbal behaviour, suffice to say there is much
disagreement by social scientists in this area.
Sequential Analysis
Sequential data analysis has been used by a number of authors (eg.,
IP: 119.92.61.209 On: Sat, 10 Oct 2015 22:04:28
Olson et al. 1994; Putnam 1991; Hawks 1987, et al) and quoted by others
(Morley and Stephenson 1977) as a complementary methodology for content
analysis. Its various techniques have been described as "methods for
dealing with behaviour as it unfolds over time..." (Sanderson 1991, in Cuomo
1994, p 171), providing meaningful insights into the behaviour (Faroane and
Dorfman 1987; Lichtenberg and Heck 1986) of both small groups (Becker-
Beck and Fische 1984) and dyads (Moran et al. 1992; Weingart et al. 1990
and 1993).
The main premise of sequential analysis is the contingency that some
serial dependency exists between events or behaviours (Moran et al. 1992;
Suen and Ary 1989). Consequently, analytical techniques include –
Dorfman 1987). Several authors (see Sanderson and Fisher 1994, for a
summary) have raised doubts as to the appropriateness of measures,
particularly in exploratory studies, suggesting the independence of
behaviours lacks a ‘common-sense perspective’, as assumed by a null
hypothesis. Indeed, communication is a social interaction of two or more
Delivered by Publishing Technology to: Chinese University of Hong Kong
Discussion
Benefits Limitations
· flexibility of research design ie., · analyses the communication
Delivered by Publishing Technology to: Chinese University of Hong Kong
Conclusion
review identifies that design of any content analytic coding mechanism needs
to consider whether data will be sequentially analysed and, therefore, the
limitations associated with the unitising of data. Careful consideration needs
to be given to the appropriateness of a design which affords a continuous
approach to data collection - this clearly gives the greatest richness to
analyses and improves potential accuracy in subsequent analyses to one
which has merely sampled data on an intermittent basis. Quite simply,
phenomena are not missed out with a continuous design. Of the two main
data unitising techniques which enable subsequent sequential analysis,
event-based offers more opportunity to analyse phenomena of an unknown
length whereas if time intervals are used, important phenomena may be
missed.
References
Tracy Harwood is a senior lecturer and teacher fellow within the Department
of Marketing at Leicester Business School, De Montfort University. Her
research specialism is in negotiation processes of strategic business
partners, on which she has worked closely with a number of major
organisations over the past 5 years and is an assistant examiner for the
Chartered Institute of Purchase and Supply. Her publications include
conference proceedings, academic journal articles, textbooks and trade
articles.