Online content analysis explained

Online content analysis or online textual analysis refers to a collection of research techniques used to describe and make inferences about online material through systematic coding and interpretation. Online content analysis is a form of content analysis for analysis of Internet-based communication.

History and definition

See main article: Content analysis.

Content analysis as a systematic examination and interpretation of communication dates back to at least the 17th century. However, it was not until the rise of the newspaper in the early 20th century that the mass production of printed material created a demand for quantitative analysis of printed words.[1]

Berelson’s (1952) definition provides an underlying basis for textual analysis as a "research technique for the objective, systematic and quantitative description of the manifest content of communication."[2] Content analysis consists of categorizing units of texts (i.e. sentences, quasi-sentences, paragraphs, documents, web pages, etc.) according to their substantive characteristics in order to construct a dataset that allows the analyst to interpret texts and draw inferences. While content analysis is often quantitative, researchers conceptualize the technique as inherently mixed methods because textual coding requires a high degree of qualitative interpretation.[3] Social scientists have used this technique to investigate research questions concerning mass media, media effects[4] and agenda setting.[5]

With the rise of online communication, content analysis techniques have been adapted and applied to internet research. As with the rise of newspapers, the proliferation of online content provides an expanded opportunity for researchers interested in content analysis. While the use of online sources presents new research problems and opportunities, the basic research procedure of online content analysis outlined by McMillan (2000) is virtually indistinguishable from content analysis using offline sources:

  1. Formulate a research question with a focus on identifying testable hypotheses that may lead to theoretical advancements.
  2. Define a sampling frame that a sample will be drawn from, and construct a sample (often called a ‘corpus’) of content to be analyzed.
  3. Develop and implement a coding scheme that can be used to categorize content in order to answer the question identified in step 1. This necessitates specifying a time period, a context unit in which content is embedded, and a coding unit which categorizes the content.
  4. Train coders to consistently implement the coding scheme and verify reliability among coders. This is a key step in ensuring replicability of the analysis.
  5. Analyze and interpret the data. Test hypotheses advanced in step 1 and draw conclusions about the content represented in the dataset.

Content analysis in internet research

Since the rise of online communication, scholars have discussed how to adapt textual analysis techniques to study web-based content. The nature of online sources necessitates particular care in many of the steps of a content analysis compared to offline sources.

While offline content such as printed text remains static once produced, online content can frequently change. The dynamic nature of online material combined with the large and increasing volume of online content can make it challenging to construct a sampling frame from which to draw a random sample. The content of a site may also differ across users, requiring careful specification of the sampling frame. Some researchers have used search engines to construct sampling frames. This technique has disadvantages because search engine results are unsystematic and non-random making them unreliable for obtaining an unbiased sample. The sampling frame issue can be circumvented by using an entire population of interest, such as tweets by particular Twitter users[6] or online archived content of certain newspapers as the sampling frame.[7] Changes to online material can make categorizing content (step 3) more challenging. Because online content can change frequently it is particularly important to note the time period over which the sample is collected. A useful step is to archive the sample content in order to prevent changes from being made.

Online content is also non-linear. Printed text has clearly delineated boundaries that can be used to identify context units (e.g., a newspaper article). The bounds of online content to be used in a sample are less easily defined. Early online content analysts often specified a ‘Web site’ as a context unit, without a clear definition of what they meant. Researchers recommend clearly and consistently defining what a ‘web page’ consists of, or reducing the size of the context unit to a feature on a website. Researchers have also made use of more discrete units of online communication such as web comments[8] or tweets.

King (2008) used an ontology of terms trained from many thousands of pre-classified documents to analyse the subject matter of a number of search engines.[9]

Automatic content analysis

See also: Document classification and Natural language processing.

The rise of online content has dramatically increased the amount of digital text that can be used in research. The quantity of text available has motivated methodological innovations in order to make sense of textual datasets that are too large to be practically hand-coded as had been the conventional methodological practice. Advances in methodology together with the increasing capacity and decreasing expense of computation has allowed researchers to use techniques that were previously unavailable to analyze large sets of textual content.

Automatic content analysis represents a slight departure from McMillan's online content analysis procedure in that human coders are being supplemented by a computational method, and some of these methods do not require categories to be defined in advanced. Quantitative textual analysis models often employ 'bag of words' methods that remove word ordering, delete words that are very common and very uncommon, and simplify words through lemmatisation or stemming that reduces the dimensionality of the text by reducing complex words to their root word.[10] While these methods are fundamentally reductionist in the way they interpret text, they can be very useful if they are correctly applied and validated.

Grimmer and Stewart (2013) identify two main categories of automatic textual analysis: supervised and unsupervised methods.Supervised methods involve creating a coding scheme and manually coding a sub-sample of the documents that the researcher wants to analyze. Ideally, the sub-sample, called a 'training set' is representative of the sample as a whole. The coded training set is then used to 'teach' an algorithm how the words in the documents correspond to each coding category. The algorithm can be applied to automatically analyze the remained of the documents in the corpus.

Unsupervised methods can be used when a set of categories for coding cannot be well-defined prior to analysis. Unlike supervised methods, human coders are not required to train the algorithm. One key choice for researchers when applying unsupervised methods is selecting the number of categories to sort documents into rather than defining what the categories are in advance.

Validation

Results of supervised methods can be validated by drawing a distinct sub-sample of the corpus, called a 'validation set'. Documents in the validation set can be hand-coded and compared to the automatic coding output to evaluate how well the algorithm replicated human coding. This comparison can take the form of inter-coder reliability scores like those used to validate the consistency of human coders in traditional textual analysis.

Validation of unsupervised methods can be carried out in several ways.

Challenges in online textual analysis

Despite the continuous evolution of text-analysis in the social science, there are still some unsolved methodological concerns. This is a (non-exclusive) list with some of this concerns:

See also

Notes and References

  1. Book: Krippendorff. Klaus. Content Analysis: An introduction to its methodology. 2012. Sage. Thousand Oaks, CA. 1.
  2. McMillan. Sally J.. The Microscope and the Moving Target: The Challenge of Applying Content Analysis to the World Wide Web. Journalism and Mass Communication Quarterly. March 2000. 77. 1. 80–98. 10.1177/107769900007700107. 143760798. 2.
  3. Book: van Selm. Martine. Jankowski. Nick. Content Analysis of Internet-Based Documents. 2005. Unpublished Manuscript. 3.
  4. Book: Riffe. Daniel. Lacy. Stephen. Fico. Frederick. Analyzing Media Messages: Using Quantitative Content Analysis in Research. 1998. Lawrence Erlbaum. Mahwah, New Jersey, London.
  5. Book: Baumgartner. Frank. Jones. Bryan. Agendas and Instability in American Politics. Chicago. 1993. University of Chicago Press. 9780226039534.
  6. Barberá. Pablo. Bonneau. Richard. Egan. Patrick. Jost. John. Nagler. Jonathan. Tucker. Joshua. Leaders or Followers? Measuring Political Responsiveness in the U.S. Congress Using Social Media Data.. Prepared for Delivery at the Annual Meeting of the American Political Science Association. 2014. 6.
  7. DiMaggio. Paul. Nag. Manish. Blei. David. Exploiting affinities between topic modeling and the sociological perspective on culture: Application to newspaper coverage of U.S. government arts funding. Poetics. December 2013. 41. 6. 570–606. 10.1016/j.poetic.2013.08.004.
  8. Mishne. Gilad. Glance. Natalie. Leave a reply: An analysis of weblog comments. Third Annual Conference on the Weblogging Ecosystem. 2006. 8.
  9. PhD . King . John D. . 2008 . Search Engine Content Analysis . Queensland University of Techbology.
  10. Grimmer. Justin. Stewart. Brandon. Text as Data: The Promise and Pitfalls of Automatic Content Analysis Methods for Political Texts. Political Analysis. 2013. 21. 3. 267–297. 10.1093/pan/mps028. 9. free.
  11. Collingwood, Loren and John Wilkerson. (2011). Tradeoffs in Accuracy and Efficiency in supervised Learning Methods, in The Journal of Information Technology and Politics, Paper 4.
  12. Gerber . Elisabeth . Lewis . Jeff . 2004 . Beyond the median: Voter preferences, district heterogeneity, and political representation . https://web.archive.org/web/20151001083853/http://www.sscnet.ucla.edu/polisci/faculty/lewis/pdf/gerberLewis.pdf . dead . October 1, 2015 . Journal of Political Economy . 112 . 6. 1364–83 . 10.1086/424737. 10.1.1.320.8707 . 16695697 .
  13. Slapin, Jonathan, and Sven-Oliver Proksch. 2008. A scaling model for estimating time-series party positions from texts.American Journal of Political Science 52(3):705–22.
  14. King, Gary, Robert O. Keohane, & Sidney Verba. (1994). Designing Social Inquiry: Scientific Inference in Qualitative Research. Princeton: Prince University Press.
  15. Book: Herring, Susan C.. Springer Netherlands. 2009. 978-1-4020-9788-1. 233–249. en. Jeremy. Hunsinger. 10.1007/978-1-4020-9789-8_14. International Handbook of Internet Research. Web Content Analysis: Expanding the Paradigm. 10.1.1.476.6090.
  16. Saldana Johnny. (2009). The Coding Manual for Qualitative Research. London: SAGE Publication Ltd.
  17. Chuang, Jason, John D. Wilkerson, Rebecca Weiss, Dustin Tingley, Brandon M. Stewart, Margaret E. Roberts, Forough Poursabzi-Sangdeh, Justin Grimmer, Leah Findlater, Jordan Boyd-Graber, and Jeffrey Heer. (2014). Computer-Assisted Content Analysis: Topic Models for Exploring Multiple Subjective Interpretations. Paper presented at the Conference on Neural Information Processing Systems (NIPS). Workshop on HumanPropelled Machine Learning. Montreal, Canada.