@article{Hase_2021, title={Actors (Automated Content Analysis)}, url={https://www.hope.uzh.ch/doca/article/view/1b}, DOI={10.34778/1b}, abstractNote={<p>Actors in coverage might be individuals, groups, or organizations, which are discussed, described, or quoted in the news.</p> <p>The datasets referred to in the table are described in the following paragraph:</p> <p>Benoit and Matuso (2020) uses fictional sentences (N = 5) to demonstrate how named entities and noun phrases can be identified automatically. Lind and Meltzer (2020) demonstrate the use of organic dictionaries to identify actors in German newspaper articles (2013-2017, N = 348,785). Puschmann (2019) uses four data sets to demonstrate how sentiment/tone may be analyzed by the computer. Using tweets (2016, N = 18,826), German newspaper articles (2011-2016, N = 377), Swiss newspaper articles (2007-2012, N = 21,280), and debate transcripts (1970-2017, N = 7,897), he extracts nouns and named entities from text. Lastly, Wiedemann and Niekler (2017) extract proper nouns from State of the Union speeches (1790-2017, N = 233).</p> <p><strong><em>Field of application/theoretical foundation:</em> </strong></p> <p>Related to theories of “<u>Agenda Setting</u>” and “<u>Framing</u>”, analyses might want to know how much weight is given to a specific actor, how these actors are evaluated and what perspectives and frames they might bring into the discussion how prominently.</p> <p><em><strong>References/combination with other methods of data collection:</strong></em></p> <p>Oftentimes, studies use both manual and automated content analysis to identify actors in text. This might be a useful tool to extend the lists of actors that can be found as well as to validate automated analyses. For example, Lind and Meltzer (2020) combine manual coding and dictionaries to identify the salience of women in the news.</p> <p> </p> <p>Table 1. Measurement of “Actors” using automated content analysis.</p> <div style="overflow-x: auto;"> <table width="100%"> <tbody> <tr> <td class="t"> <p><strong>Author(s)</strong></p> </td> <td class="t"> <p><strong>Sample</strong></p> </td> <td class="t"> <p><strong>Procedure</strong></p> </td> <td class="t"> <p><strong>Formal validity check with manual coding as benchmark*</strong></p> </td> <td class="t"> <p><strong>Code </strong></p> </td> </tr> <tr> <td class="t"> <p>Benoit & Matuso (2020)</p> </td> <td class="t"> <p>Fictional sentences</p> </td> <td class="t"> <p>Part-of-Speech tagging; syntactic parsing</p> </td> <td class="t"> <p>Not reported</p> </td> <td class="t"> <p><a href="https://cran.r-project.org/web/packages/spacyr/vignettes/using_spacyr.html">https://cran.r-project.org/web/packages/spacyr/vignettes/using_spacyr.html</a></p> </td> </tr> <tr> <td class="t"> <p>Lind & Meltzer</p> <p>(2020)</p> </td> <td class="t"> <p>Newspapers</p> </td> <td class="t"> <p>Dictionary approach</p> </td> <td class="t"> <p>Reported</p> </td> <td class="t"> <p><a href="https://osf.io/yqbcj/?view_only=369e2004172b43bb91a39b536970e50b">https://osf.io/yqbcj/?view_only=369e2004172b43bb91a39b536970e50b</a></p> </td> </tr> <tr> <td class="t"> <p>Puschmann (2019)</p> </td> <td class="t"> <p>(a) Tweets</p> <p>(b) German newspaper articles</p> <p>(c) Swiss newspaper articles</p> <p>(d) United Nations General Debate Transcripts</p> </td> <td class="t"> <p>Part-of-Speech tagging; syntactic parsing</p> </td> <td class="t"> <p>Not reported</p> </td> <td class="t"> <p><a href="http://inhaltsanalyse-mit-r.de/ner.html">http://inhaltsanalyse-mit-r.de/ner.html</a></p> </td> </tr> <tr> <td class="t b"> <p>Wiedemann & Niekler (2017)</p> </td> <td class="t b"> <p>State of the Union speeches</p> </td> <td class="t b"> <p>Part-of-Speech tagging</p> </td> <td class="t b"> <p>Not reported</p> </td> <td class="t b"> <p><a href="https://tm4ss.github.io/docs/Tutorial_8_NER_POS.html">https://tm4ss.github.io/docs/Tutorial_8_NER_POS.html</a></p> </td> </tr> </tbody> </table> </div> <div>*<em>Please note that many of the sources listed here are tutorials on how to conducted automated analyses – and therefore not focused on the validation of results. Readers should simply read this column as an indication in terms of which sources they can refer to if they are interested in the validation of results.</em></div> <p><strong>References</strong></p> <p>Benoit, K., & Matuso. (2020). <em>A Guide to Using spacyr</em>. Retrieved from https://cran.r-project.org/web/packages/spacyr/vignettes/using_spacyr.html</p> <p>Lind, F., & Meltzer, C. E. (2020). Now you see me, now you don’t: Applying automated content analysis to track migrant women’s salience in German news. <em>Feminist Media Studies</em>, 1–18.</p> <p>Puschmann, C. (2019). <em>Automatisierte Inhaltsanalyse mit R</em>. Retrieved from http://inhaltsanalyse-mit-r.de/index.html</p> <p>Wiedemann, G., Niekler, A. (2017). <em>Hands-on: a five day text mining course for humanists and social scientists in R</em>. Proceedings of the 1st Workshop Teaching NLP for Digital Humanities (Teach4DH@GSCL 2017), Berlin. Retrieved from https://tm4ss.github.io/docs/index.html</p>}, journal={DOCA - Database of Variables for Content Analysis}, author={Hase, Valerie}, year={2021}, month={Mar.} }