Blog

Workshop: Scientific Data Visualization

Join us in the upcoming workshop on Scientific Data Visualization, beneficial for researchers, students, and professionals interested in improving their skills in quantitative data visualization. On December 4, 2025, Ansgar Hudde, lecturer in Sociology at the University of Cologne, will lead an online, hands-on session introducing strategies for creating clear and effective visualizations in Stata.

The workshop will cover key aspects of graph design, including working with text, colours, axes, reference lines, transparency, fonts, and number formatting. Through interactive exercises with example datasets, participants will gain practical experience applying these principles and build confidence in producing polished, informative figures. By the end of the session, they will be able to create and refine visualizations that strengthen their empirical research and academic publications.

For more details, please visit our program page. We hope to see you there! 

Introducing Overleaf at WI

Writing is at the core of scientific work: papers, research proposals, reports, and reviews are some of the most important elements of scientists’ everyday tasks. Although the act of writing is a complex subject itself, the tools that enable it are manifold and affect the efficiency of any research project.

Google Docs has established itself as a helpful tool for scientific writing over the past years. It is reminiscent of the well-known Microsoft Word, and is easy-to-use and convenient. As a web service, it does not require installing dedicated software, but works in any modern browser. Its main feature is collaborative writing – documents can be shared either via email address or a link, and edited together in real time, also allowing to post comments and track changes.

At the same time, LaTeX has grown in popularity as an alternative typesetting system and markup language for scientific and technical writing. As opposed to software like Microsoft Word and Google Docs, it distinguishes between the source of a text document which consists of the content as well as formatting instructions, and the rendered document (e.g., in .pdf format) which presents the formatted content accordingly. This separation between source and result enables more precision, since content as well as formatting are both provided in plain text, without applying the formatting right away. There are different LaTeX editors, such as TeXstudio or TeXMaker, that allow to create and edit documents. However, especially in reproducible research, LaTeX is popular for its compatibility with programming frameworks and libraries such as R Markdown, Quarto, Sweave, or PyLaTeX. These allow users to merge text and formatting with information in the programming environment (e.g., analysis results, tables, or figures) using placeholders that are automatically replaced with the current information whenever the document is rendered.

Overleaf is a tool for text editing using LaTeX, but featuring the advantages of Google Docs: It does not require installing separate software, as it is accessible through a web browser, and it enables collaborative writing. While Overleaf can be used on the official website either for free (with limited features) or with a subscription plan (to unlock more), it is also possible to host an instance locally.

To meet the growing demand at WI, avoid subscription costs, and have full control over stored and shared data, we hereby present the brand new Overleaf instance at the Weizenbaum Institute. It not only allows WI researchers, but also their external colleagues to work on text documents together in real time. Overleaf (Community Edition) is hosted by the IT department and administered by the Methods Lab, and it can be accessed at https://overleaf.weizenbaum-institut.de. To use it, an account is required, which can be requested from the Methods Lab. For further instructions, please consult the Overleaf page in the internal Wiki at WI.

Happy writing!

Spotlight: Analyzing Digital Aesthetics—AR Filters and the Commodification of Identity

The tendency to measure our worth against unrealistic beauty standards is hardly a new concern, yet it persists even as feminist theory challenges these ideals and media efforts claim to represent diverse bodies. Platforms like TikTok, for example, argue their core values prioritize diversity, with policies that uplift underrepresented communities. Still, self-objectification of both women and men is largely driven by mainstream media (Rodgers, 2015), significantly impacting users’ self-esteem. An increasing body of research emphasizes self-objectification in reported decline of cognitive functioning and is associated with many mental health issues such as depression and eating disorders (Fredrickson and Roberts, 1997). But with such widespread awareness, why does self-objectification remain so prevalent? 

Researchers Corinna Canali and Miriam Doh uncover the “truth” behind the algorithmic structures that construct identities and perpetuate this self-objectification online. Bodies and forms of self-expression deemed as “true,” “normal,” or “obscene,” are not actually objective categories. They are rather social constructs created by institutions who have the power to define them, such as governments, corporations, or tech platforms. One structure that reinforces self-objectification, and distorts self-perception is the use of augmented reality filters. These AR filters have become popular across social media networks that an estimated 90 percent of American young adults visit daily, with 58 percent of teens using TikTok alone (Bhandari & Bimo, 2022). The ways in which these filters, incentivized by algorithmic systems, affect identity, prefer uniformity, and marginalize diversity are explored in Filters of Identity: AR Beauty and the Algorithmic Politics of the Digital Body, co-authored by Corinna and Mariam Doh.

Corinna is a research associate at the Weizenbaum Institute and Design Research Lab, and her project partner, Mariam, is a PhD student with the AI for the Common Good Institute / Machine Learning Group at the Université Libre de Bruxelles. Corinna’s initial inspiration began with a study on nudity in digital and social media contexts. This led to a broader analysis of digital governance engrained “obscenifying” beliefs, where systems are built on the assumption that certain bodies, identities, and forms of expression are inherently obscene. From this, she has developed extensive experience across institutions in examining how platforms and policies act as regulatory infrastructures that shape who is seen, censored, or silenced in digital environments.

In an interview conducted for this article, Corinna discusses how AI-driven beauty filters do more than simply mirror existing beauty norms, but actively construct and reinforce them.

Although TikTok’s effect house policy has been implemented to ban filters that promote unrealistic beauty standards, the platform continues to circulate effects that do exactly that. As an example, Corinna analyzes the Bold Glamour filter, which edits users’ faces in ways that are difficult to detect but tend to favor Eurocentric and cisgender features such as lighter skin, narrow noses, and larger eyes. It also frequently misidentifies or distorts the facial features of certain racial groups. The highest “misclassification rates” were detected in black women, where the filter failed to accurately detect a person’s face or altered their features in ways that don’t reflect their real appearance. This systematic exclusion of non-conforming bodies manifests into a larger trend in technological design, where digital systems tend to perpetuate dominant norms. Bodies that fall outside these norms, including those that are disabled, fat, Indigenous, or Black, are frequently invisible or misrepresented within the algorithm. 

Corinna argues that these marginalized identities also have the most underrepresented narratives when it comes to bias in content moderation:

“Also, if you (tech company) own all the means of knowledge, circulation and production it’s quite difficult to allow any other narrative to exist within these spaces.” 

She then identifies a serious problem in the subtle ways this discrimination persists unnoticed outside academic circles. She noted that while users may recognize individual instances of bias, they often lack the tools or social power to challenge them within digital spaces.

Notably, the marginalization of these groups has deep roots in the broader historical context of racial capitalism, a system in which racial hierarchies are used to justify and sustain economic exploitation (Ralph & Singhal, 2019). When asked about whether these systems were just reflecting the norms obtained from biased datasets, or if they are actively structured to serve deeper capitalist goals, Corinna describes the influence of deliberate human intervention:

“The algorithm does not work on its own. It works within a system that is constantly being tweaked by real-time content moderation from humans.” 

She underlines the active decisions made by people moderating content from policies servicing business and corporate models, ultimately benefitting the platform’s profitability. 


The first row shows how the filter incorrectly applies a female-targeted transformation to a male face, while the second row shows a female face altered with a male-targeted filter.

Research shows that frequent use of appearance-enhancing filters is associated with increased body image concerns and higher levels of body dissatisfaction (Caravelli et al., 2025). This lowered self-esteem carries significant economic consequences, with body dissatisfaction-related issues costing an estimated $226 and $507 billion globally in 2019 alone (ibid.). Such widespread dissatisfaction is not incidental but is embedded within the algorithmic logics of these platforms.

To elaborate, these filters produce idealized versions of the self that adhere to capitalist interests by making users more desirable, and easier to manipulate with ads. One’s face becomes a data point, and their emotions and desires are recognized as commodities. Additional findings emphasize how these automated categorization systems and targeted advertisements actively shape user identity. According to Bhandari and Bimo (2022), the system influences how people construct, perform, and manage their sense of self in ways that benefit the platform economically but may be harmful to the user.

Corinna addresses this when asked whether TikTok truly allows space for authentic self-expression, or if users are instead shaping their identities to align with platform incentives. She feels conflicted, but responds with insight into how the platform promotes self-governing systems, where the primary commodities are users, their data, and their attention:

“When there is this constant negotiation between your own self expression and agency against this corporate logic of profit, it is always difficult to understand to what extent you are actually free to be who you want to be.”

To promote a more authentic version of self-expression, Corinna’s work suggests filters that give inclusive filtering options, allowing users to specifically choose their own aesthetic preferences. However, the persistence of algorithmic bias reveals the challenge of resisting systemic power: even when AR filters are redesigned to reject dominant beauty standards, the algorithm may continue to privilege a certain aesthetic.

Bearing this in mind, Corinna introduces a transdisciplinary approach utilizing design and theory. She describes the nature of this problem as multifaceted and interconnected, necessitating an equally dynamic solution that addresses the whole system. She describes the benefits of design, as her background in visual cultures allows her to embrace visual ethnographic methods to interpret the implications of images for a unique perspective on bias. She says,

“Images are a product of visual tradition that has its own biases outside of technology. Technology has inherited a lot of this bias, discrimination and representation from other traditions that live outside of it.” 

In this way, Corinna uses design as a tool to reveal and critique bias. This approach does not require technical expertise, but instead encourages people to recognize the significance of what they’re looking at, and in doing so, begin to see the discriminatory practices embedded within it.

Corinna’s next steps address both the internal complexity of these systems and the broader socio-political contexts in which they function. She offers one possible solution in creating opportunities to rethink and rebuild these platforms, but for this, there needs to be space for alternative systems. Corinna highlights that it is extremely difficult now to access any social network that does not come from big tech companies, because they have created conditions where they hold a near-monopoly over these platforms and infrastructures. The process will require time and multiple layers of reworking, starting with how users are educated about these technologies.

To address the lack of adequate understanding on these issues, she plans to develop tools that promote media literacy by helping users acknowledge what these systems are, how they function, and what their greater societal impacts may be. This perspective would stress the reciprocal relationship between society and technology, emphasizing how each continuously shapes the other.

The Design Research Lab and Berlin Open Lab have many projects that explore research through design, uniquely combining critical theory from the humanities with hands-on material production. The lab takes a different path from traditional product design by emphasizing critical reflection on both how products are made and the ways they shape society. Several research projects in this lab consider this approach, including one that addresses the theme of surveillance capitalism discussed here. For a deeper look into how the lab works, check out the spotlight article linked here!

New publication: Bursting Self-reports? Comparing Sampling Frequency Effects of Mobile Experience Sampling Method on Compliance, Attrition, and Sample Biases

The Methods Lab is excited to share a new publication in Journal of Quantitative Description: Digital Media, authored by researchers Jakob Ohme, Timothy Charlton, Roland Toth, Theo Araujo, and Claes H. de Vreese. This paper, titled Bursting Self-reports? Comparing Sampling Frequency Effects of Mobile Experience Sampling Method on Compliance, Attrition, and Sample Biases, explores the effects of different sampling frequencies in experience sampling studies on compliance, sample biases, and reactivity of measures in the context of digital media use.

The study presents an analysis of the impact of daily-intensive burst measures (seven surveys per day) versus hourly-intensive burst measures (12 surveys over two hours per day) on compliance, attrition, and sample biases. To do so, the authors used data collected through the Mobile Experience Sampling Method (MESM) with a mobile-only sample of Dutch Internet users. They reveal differences in compliance, attrition, and sample biases between the two sampling frequency designs.

The results indicate general problems with MESM studies, such as sample attrition during onboarding and sample biases due to age, education, tech savviness, and privacy literacy. They also show that hourly-intensive burst measures lead to lower MESM protocol compliance compared to a more spread-out sampling schedule. However, the study also finds that the average response rate across seven study days does not strongly decrease, suggesting that while the number of measurements may be lower using hourly-intensive burst measures, the day-to-day decrease is a minor issue.

The study highlights the importance of considering systematic biases in MESM studies, particularly during the recruitment phase, and suggests that researchers should be transparent and mindful about these. The findings have implications for future ESM studies with high-frequency sampling, suggesting that researchers should carefully consider the sampling frequency design to minimize biases and ensure high-quality data.

New Publication: A Mixed Methods Study of Smartphone Use

The Methods Lab is excited to share a new publication in the Journal of Quantitative Description: Digital Media, authored by data scientist Roland Toth along with researchers Douglas Parry, and Martin Emmer from the Weizenbaum institute. This paper, titled From Screen Time to Daily Rhythms: A Mixed Methods Study of Smartphone Use Among German Adults, explores how much, when, how, and under which circumstances Germans use their smartphones throughout the day. 

The study presents a detailed analysis of smartphone usage characteristics, showing engagement in aggregate as well as how it varies throughout the day, and associations with the user’s socio-demographic characteristics. To do so, the authors used a mix of the Mobile Experience Sampling Method (MESM), Android event logs, and iOS data donations. They reveal distinct temporal patterns that are beneficial for understanding the broader contexts, motivations, and situational factors shaping different types of mobile interactions. It also outlines practical implications for researchers employing longitudinal and real-time measurement methods in interdisciplinary and social science research. This comprehensive analysis provides a strong basis for further exploration of the psychological, social, and behavioral dimensions of smartphone use.

Tool presentation: ChatAI (Academic Cloud)

Most commercial, browser-based generative AI tools have usage or licensing restrictions that prevent users from exploring their full potential. All members of the Weizenbaum Institute now have access to a new chatbot providing unrestricted use of generative AI, thanks to an institution in Göttingen that locally hosts a range of large language models under the name ChatAI. The platform is essentially like ChatGPT, but it is hosted in Germany, and can be used free of charge without licensing costs. It also includes other useful tools, such as ImageAI, which generates images from text prompts, and VoiceAI, which can interpret human speech.

To get started, simply visit the Academic Cloud website and click on Login in the top right corner. On the login page, choose Federated Login on the right-hand side, then select Weizenbaum Institute from the list of institutions. From there, you can log in using your WI account credentials and follow the on-screen instructions.

After logging in, open the ChatAI tool in the list of services. Within ChatAI, the model selector at the top center of the interface allows you to choose from different language models depending on your needs. Click the ChatAI logo in the top left corner to easily switch between different tools to generate text, images, or work with audio input. With unlimited access to these tools, you can find efficient new ways to enhance your research and creative projects.

Workshop: Finding Frames with RoBERTa – A Crash Course

The Methods Lab is pleased to host a hands-on workshop led by Dr. Vihang Jumle (University of Bern) on automating frame analysis using RoBERTa. This practical session teaches social science researchers how to apply pre-trained language models to scale text coding – transforming manual content analysis into a fast, reproducible process. Participants will learn to fine-tune models, preprocess data, apply data augmentation, and evaluate results using precision, recall, and cross-validation – using their own research datasets. Designed for intermediate Python users, the workshop emphasizes real-world application and project-based learning. Ideal for researchers in communication studies, political science, and sociology. Register now and bring your data to automate your next analysis!

To learn more, please visit the program page. We hope to see you there!

Research Stay at Dartmouth College

This spring, the Methods Lab student assistant, Diana Ignatovich, spent four months at Dartmouth college to research octopus cognition and visual processing.

Hidden in the New England wilderness is an underground laboratory housing three male Octopus bimaculoides that were shipped from the Pacific Ocean for non-invasive study using underwater electroencephalography (EEG). One of the octopuses was named Joseph—after the Weizenbaum Institute’s namesake, computer scientist Joseph Weizenbaum.

Rather than euthanizing the octopus or holding its tentacles down to understand its incredibly complex and decentralized nervous system, the underwater EEG apparatus used here is methodologically unique, to ensure no octopuses were harmed. This method records the brain’s electrical activity by detecting signals from groups of neurons, which are amplified by the EEG machine and studied as brain waves. To record this data, the experimental tank consisted of a clear plastic cube with two printed circuit boards on top and bottom, lined with tripolar concentric ring electrodes and submerged in saltwater. The tanks were also enriched with toys to encourage cognitive stimulation, and all experimental procedures were conducted in accordance with ethical guidelines approved by the Institutional Animal Care and Use Committee (IACUC).

The primary project for this stay was in assessing the neural critical flicker fusion frequency (CFF) of the octopus via steady-state visual evoked potentials of the EEG power spectrum. The CFF threshold refers to the point at which a rapidly flickering light is perceived as steady, which indicates the speed of visual information processing in the brain. This gives insight for neural and visual processing efficiency as well as cognitive functions such as attentional control and overall responsiveness to changing environments. It is also commonly measured in psychophysics via behavioral paradigms, where a participant indicates the observed flicker fusion. However, since Joseph could not tell us about this boundary, the threshold was marked by a drop in EEG signal amplitude as flicker perception diminished. These experiments, performed using LED light at different brightness levels, were then repeated with human testing for comparison, to determine whether octopuses are better adapted to low-light conditions due to their underwater habitat.

The octopuses demonstrated higher critical flicker fusion frequency thresholds compared to humans, likely due to their evolutionary history and environmental conditions, where quick visual responses are necessary to spot prey or avoid predators.

It’s extraordinary that octopuses, despite lacking the rod and cone photoreceptor cells found in most eyes, have adapted to match or even exceed the critical flicker fusion thresholds of creatures like humans. These results are an exciting glimpse into the diversity and complexity of how other species experience their own vastly different worlds in nature.

Reference: “When the Field is Online” Newsletter

When the Field Is Online is a monthly newsletter by Janet Salmons, PhD, qualitative methodologist, and author of 12 books in academic writing and research collaboration. Building on this extensive experience, Janet introduces thoughtful and creative strategies to overcome methodological challenges, and to connect meaningfully with others in an increasingly digital research landscape.

Her latest topics have reviewed virtual focus groups, ethics in remote research, reflexivity in data analysis, and tips for recruiting authentic voices. Each issue is unique, with a combination of original essays, thoughtful analysis, instructional videos, open access resources, and a bonus of her own self-drawn illustrations. The blog also links other relevant content and reflections on research obstacles from designing effective studies to asking stronger research questions. For anybody interested in expanding their qualitative research skills, strengthening digital communication, or discovering new ways to connect virtually, When the Field Is Online offers thorough guidance and novel ideas to inspire your work.

Most of this newsletter is freely available, but to access the full content, please sign up here to become a paid subscriber!

Computing Resources for WI

As digital methods advance rapidly, quantitative empirical research requires greater computing power. This includes complex statistical analyses, model training, and operating generative AI models. The necessary hardware is expensive and challenging to maintain, particularly at the institutional level (e.g., due to high temperatures in HPC clusters). Since not every institution can (or should) set up such hardware independently, resources are shared through collaborations with other institutions.

Against this background, the Weizenbaum Institute is exploring options to enable both low-demand computing tasks (e.g., a virtual machine for background web scraping) and high-demand tasks (e.g., running current large language models) for its researchers. Recently, the following options have been identified (some of which are only available to Berlin scientists):

Datenzentrum Berlin (Berlin University Alliance)

  • Not established yet, but planned

de.NBI Cloud (Deutsches Kompetenzzentrum Clouddienste)

  • free to use with scientific project proposal: https://datenkompetenz.cloud/
  • “Self service” virtual machines
  • currently trying to support digital humanities and social science

FUB-IT (Freie Universität Berlin)

  • Free to use (FU account required; guest accounts are available)
  • High-performance cluster (HPC)
  • Software partially included, some can be requested

Gauss Computing Center

  • Up to very large scale projects (both HPC and AI) on proposal in different calls: https://www.gauss-centre.eu/for-users/hpc-access
  • Several equipment up to supercomputers SuperMUC I and II usable

Gesellschaft für wissenschaftliche Datenverarbeitung mbH Göttingen (GWDG):

Galaxy Server

  • Community-driven web-based analysis platform for life science research
  • Free to use batch system with many tools usable by browser, huge data sizes are no problem
  • Although coming from Bioinformatics, there are many tools for text processing, audio and image analyses, as well as statistics can be put together in pipelines.

HPI (Hasso Plattner Institut) UP Potsdam

Zentraleinrichtung Campusmanagement (TU Berlin)

  • Free to use (TUB account required)
  • High-performance cluster
  • Software partially included, some can be requested

Zuse Institute Berlin (ZIB)

  • Support from the state of Berlin, linked to BUA, e.g. BI-FOLD also calculates there
  • Provides tailor-made solutions, e.g. can make own instances of HAWKI, LibreChat or similar available with WI licenses
  • Resources available only by collaboration.

Apart from these external offers, the Weizenbaum Institute itself already provides some services to its researchers:

  • Virtual machines
  • Jupyterhub
  • Gitlab

Researchers who are interested in using any of these services can follow the instructions on the according websites for external services. For internal services, instructions are provided in the internal WI Wiki.

Disclaimer: This list will be updated regularly. If you know of any other resources that are available to members of the Weizenbaum Institute, please let us know.