Publikation: Computer Vision for Protest Analysis
Dateien
Datum
Autor:innen
Herausgeber:innen
ISSN der Zeitschrift
Electronic ISSN
ISBN
Bibliografische Daten
Verlag
Schriftenreihe
Auflagebezeichnung
URI (zitierfähiger Link)
Internationale Patentnummer
Link zur Lizenz
Angaben zur Forschungsförderung
Projekt
Open Access-Veröffentlichung
Core Facility der Universität Konstanz
Titel in einer weiteren Sprache
Publikationstyp
Publikationsstatus
Erschienen in
Zusammenfassung
How can computer vision help us to understand protests better? Every day, people take to the streets to protest, and images of these events are shared thousands of times on social media. While qualitative studies have effectively demonstrated that protests are diverse and highly dynamic, quantitative research faces the challenge of capturing this nuanced information. However, protest images offer a unique opportunity to do so, as each image provides detailed documentation of what is happening at a particular time and place. Since these images are shared thousands of times on protest days, they can be used to reconstruct the events as they unfold. Researchers have rarely analyzed these images due to the difficulty of extracting protest-related information from them. Fortunately, recent advances in computer vision are changing this landscape. Computers are now capable of performing many visual tasks, including extracting high-level insights from images and videos. Dedicated models have already been trained to recognize protest images and assess the level of violence depicted in them. Additionally, many generic models can be adapted from computer science to applications in social sciences. For instance, segmentation models can identify a wide range of objects in images, such as people and faces. Although these tasks could theoretically be performed manually, the large scale of images on social media renders this infeasible. Therefore, researchers increasingly rely on computer vision methods to efficiently extract information from these images. This dissertation explores different applications in which computer vision enhances our understanding of protests. To achieve this, readily available computer vision methods are adopted, trained, and optimized specifically for analyzing protest images. These methods facilitate the extraction of various characteristics from these images, enabling a deeper analysis of the protests themselves. A distinct image dataset complements each method. The first dataset comprises more than 140,000 images collected from social media, with annotations indicating whether each image depicts a protest or not. This dataset aims to provide a comprehensive overview across ten different countries. The second dataset focuses on capturing protest periods in specific cities, covering 13 protest episodes and incorporating approximately 22,000 images. The findings reveal that persons, flags, and signboards are important objects in protest images. But particular features of protests vary across different countries and protest episodes. The results also indicate that the escalation of protest events can be tracked through images shared on social media, allowing for predictions of protest dynamics on the same day. However, predictions for the following day show only marginal improvements. Experimental results highlight how individuals perceive protests through sequences of images. If generative computer vision models manipulate crowds in these protest images, it threatens public perception, as estimates of crowd sizes become distorted. Overall, these findings expand our understanding of protests in a world saturated with visual information, opening exciting avenues for future research in protest studies and other fields of social science.
Zusammenfassung in einer weiteren Sprache
Fachgebiet (DDC)
Schlagwörter
Konferenz
Rezension
Zitieren
ISO 690
SCHOLZ, Stefan, 2025. Computer Vision for Protest Analysis [Dissertation]. Konstanz: Universität KonstanzBibTex
@phdthesis{Scholz2025Compu-75457,
title={Computer Vision for Protest Analysis},
year={2025},
author={Scholz, Stefan},
address={Konstanz},
school={Universität Konstanz}
}RDF
<rdf:RDF
xmlns:dcterms="http://purl.org/dc/terms/"
xmlns:dc="http://purl.org/dc/elements/1.1/"
xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
xmlns:bibo="http://purl.org/ontology/bibo/"
xmlns:dspace="http://digital-repositories.org/ontologies/dspace/0.1.0#"
xmlns:foaf="http://xmlns.com/foaf/0.1/"
xmlns:void="http://rdfs.org/ns/void#"
xmlns:xsd="http://www.w3.org/2001/XMLSchema#" >
<rdf:Description rdf:about="https://kops.uni-konstanz.de/server/rdf/resource/123456789/75457">
<dcterms:isPartOf rdf:resource="https://kops.uni-konstanz.de/server/rdf/resource/123456789/42"/>
<dcterms:title>Computer Vision for Protest Analysis</dcterms:title>
<dcterms:abstract>How can computer vision help us to understand protests better? Every day, people take to the streets to protest, and images of these events are shared thousands of times on social media. While qualitative studies have effectively demonstrated that protests are diverse and highly dynamic, quantitative research faces the challenge of capturing this nuanced information. However, protest images offer a unique opportunity to do so, as each image provides detailed documentation of what is happening at a particular time and place. Since these images are shared thousands of times on protest days, they can be used to reconstruct the events as they unfold. Researchers have rarely analyzed these images due to the difficulty of extracting protest-related information from them. Fortunately, recent advances in computer vision are changing this landscape. Computers are now capable of performing many visual tasks, including extracting high-level insights from images and videos. Dedicated models have already been trained to recognize protest images and assess the level of violence depicted in them. Additionally, many generic models can be adapted from computer science to applications in social sciences. For instance, segmentation models can identify a wide range of objects in images, such as people and faces. Although these tasks could theoretically be performed manually, the large scale of images on social media renders this infeasible. Therefore, researchers increasingly rely on computer vision methods to efficiently extract information from these images. This dissertation explores different applications in which computer vision enhances our understanding of protests. To achieve this, readily available computer vision methods are adopted, trained, and optimized specifically for analyzing protest images. These methods facilitate the extraction of various characteristics from these images, enabling a deeper analysis of the protests themselves. A distinct image dataset complements each method. The first dataset comprises more than 140,000 images collected from social media, with annotations indicating whether each image depicts a protest or not. This dataset aims to provide a comprehensive overview across ten different countries. The second dataset focuses on capturing protest periods in specific cities, covering 13 protest episodes and incorporating approximately 22,000 images. The findings reveal that persons, flags, and signboards are important objects in protest images. But particular features of protests vary across different countries and protest episodes. The results also indicate that the escalation of protest events can be tracked through images shared on social media, allowing for predictions of protest dynamics on the same day. However, predictions for the following day show only marginal improvements. Experimental results highlight how individuals perceive protests through sequences of images. If generative computer vision models manipulate crowds in these protest images, it threatens public perception, as estimates of crowd sizes become distorted. Overall, these findings expand our understanding of protests in a world saturated with visual information, opening exciting avenues for future research in protest studies and other fields of social science.</dcterms:abstract>
<dc:date rdf:datatype="http://www.w3.org/2001/XMLSchema#dateTime">2025-12-11T13:00:01Z</dc:date>
<dcterms:rights rdf:resource="http://creativecommons.org/licenses/by/4.0/"/>
<dcterms:hasPart rdf:resource="https://kops.uni-konstanz.de/bitstream/123456789/75457/4/Scholz_2-1qouwbri54guc8.pdf"/>
<dc:creator>Scholz, Stefan</dc:creator>
<dc:rights>Attribution 4.0 International</dc:rights>
<void:sparqlEndpoint rdf:resource="http://localhost/fuseki/dspace/sparql"/>
<dc:contributor>Scholz, Stefan</dc:contributor>
<dc:language>eng</dc:language>
<dspace:hasBitstream rdf:resource="https://kops.uni-konstanz.de/bitstream/123456789/75457/4/Scholz_2-1qouwbri54guc8.pdf"/>
<bibo:uri rdf:resource="https://kops.uni-konstanz.de/handle/123456789/75457"/>
<dspace:isPartOfCollection rdf:resource="https://kops.uni-konstanz.de/server/rdf/resource/123456789/42"/>
<dcterms:issued>2025</dcterms:issued>
<dcterms:available rdf:datatype="http://www.w3.org/2001/XMLSchema#dateTime">2025-12-11T13:00:01Z</dcterms:available>
<foaf:homepage rdf:resource="http://localhost:8080/"/>
</rdf:Description>
</rdf:RDF>