Join us in the upcoming workshop on Scientific Data Visualization, beneficial for researchers, students, and professionals interested in improving their skills in quantitative data visualization. On December 4, 2025, Ansgar Hudde, lecturer in Sociology at the University of Cologne, will lead an online, hands-on session introducing strategies for creating clear and effective visualizations in Stata.
The workshop will cover key aspects of graph design, including working with text, colours, axes, reference lines, transparency, fonts, and number formatting. Through interactive exercises with example datasets, participants will gain practical experience applying these principles and build confidence in producing polished, informative figures. By the end of the session, they will be able to create and refine visualizations that strengthen their empirical research and academic publications.
For more details, please visit our program page. We hope to see you there!
Writing is at the core of scientific work: papers, research proposals, reports, and reviews are some of the most important elements of scientists’ everyday tasks. Although the act of writing is a complex subject itself, the tools that enable it are manifold and affect the efficiency of any research project.
Google Docs has established itself as a helpful tool for scientific writing over the past years. It is reminiscent of the well-known Microsoft Word, and is easy-to-use and convenient. As a web service, it does not require installing dedicated software, but works in any modern browser. Its main feature is collaborative writing – documents can be shared either via email address or a link, and edited together in real time, also allowing to post comments and track changes.
At the same time, LaTeX has grown in popularity as an alternative typesetting system and markup language for scientific and technical writing. As opposed to software like Microsoft Word and Google Docs, it distinguishes between the source of a text document which consists of the content as well as formatting instructions, and the rendered document (e.g., in .pdf format) which presents the formatted content accordingly. This separation between source and result enables more precision, since content as well as formatting are both provided in plain text, without applying the formatting right away. There are different LaTeX editors, such as TeXstudio or TeXMaker, that allow to create and edit documents. However, especially in reproducible research, LaTeX is popular for its compatibility with programming frameworks and libraries such as R Markdown, Quarto, Sweave, or PyLaTeX. These allow users to merge text and formatting with information in the programming environment (e.g., analysis results, tables, or figures) using placeholders that are automatically replaced with the current information whenever the document is rendered.
Overleaf is a tool for text editing using LaTeX, but featuring the advantages of Google Docs: It does not require installing separate software, as it is accessible through a web browser, and it enables collaborative writing. While Overleaf can be used on the official website either for free (with limited features) or with a subscription plan (to unlock more), it is also possible to host an instance locally.
To meet the growing demandat WI, avoid subscription costs, and have full control over stored and shared data, we hereby present the brand new Overleaf instance at the Weizenbaum Institute. It not only allows WI researchers, but also their external colleagues to work on text documents together in real time. Overleaf (Community Edition) is hosted by the IT department and administered by the Methods Lab, and it can be accessed at https://overleaf.weizenbaum-institut.de. To use it, an account is required, which can be requested from the Methods Lab. For further instructions, please consult the Overleaf page in the internal Wiki at WI.
The Methods Lab is pleased to host a hands-on workshop led by Dr. Vihang Jumle (University of Bern) on automating frame analysis using RoBERTa. This practical session teaches social science researchers how to apply pre-trained language models to scale text coding – transforming manual content analysis into a fast, reproducible process. Participants will learn to fine-tune models, preprocess data, apply data augmentation, and evaluate results using precision, recall, and cross-validation – using their own research datasets. Designed for intermediate Python users, the workshop emphasizes real-world application and project-based learning. Ideal for researchers in communication studies, political science, and sociology. Register now and bring your data to automate your next analysis!
To learn more, please visit the program page. We hope to see you there!
Join us for the workshop Introduction to MAXQDA, designed for all researchers, students, and professionals interested in qualitative data analysis. On May 28th, 2025, at the Weizenbaum Institute, certified MAXQDA trainer Dr. phil. Aikokul Maksutova will lead a basic yet comprehensive workshop introducing the software’s core features, aligning with the key stages of digital qualitative research.
This event will offer guidance on MAXQDA’s essential tools for documenting, coding, and analyzing qualitative data. Participants will become familiar with navigating the Code System and a range of additional features, such as functions for exporting data, linking memos, and generating visualizations. Each segment will include hands-on activities using various datasets, enabling participants to confidently apply the skills they’ve learned on their own.
To conclude, special guest and representative of MAXQDA, Ms. Tamara Pataki, will inform participants of the software’s latest innovations and host an open Q&A session.
To learn more, please visit our program page. We hope to see you there!
In a joint effort, the Career Development and the Methods Lab are excited to announce the hybrid “Career Tutorial on LLMs for all Expertise Levels”. In this tutorial, beginning with fundamental concepts of LLMs and in-context learning, we’ll address the “Needle in the Haystack Problem” and compare ultra-long context models with RAG approaches. Through practical demonstrations, participants will gain hands-on experience with RAG’s core functionalities and understand its objectives. The session delves into scaling solutions using vector databases and advanced implementations, including chunking strategies, hybrid RAG, and graph-based RAG architectures. We conclude with an overview of emerging trends, examining agentic RAG and the integration of reasoning models in deep research applications. This comprehensive exploration equips attendees with both theoretical knowledge and practical insights into the latest developments in AI language models.
For more information, visit our program page. We are looking forward to your participation!
On February 6th, 2025, LK Seiling facilitated a workshop for an Introduction to Git, with support from Sascha Kostadinoski and Quentin Bukold. This was co-organized by the Methods Lab and took place at the Weizenbaum institute. The hybrid event provided a thorough overview on the foundation of Git and its relative platforms for about 30 participants.
Firstly, Git was introduced for its general relevance. Seiling explored the qualities of its version control system and the advantages of efficiently managing changes to files. Its widespread use and accessibility were also highlighted by the software’s free and open source application. At its core, Git enables collaborative work by allowing concurrent adjustments to files by multiple participants and offers a system to track the changes made without requiring alterations to the original file.
Next, participants were invited to open the Terminal and guided through some basic commands. To this end, commands for traversing directories, creating, moving, organizing, and deleting files were explained and demonstrated in detail.
LK Seiling explains how to stage and commit changes
In the second hour of the workshop, Seiling encouraged participants to implement these basics by imagining the context of a classic Python project, one that might require collaborative engagement. Here, Python scripts were saved, renamed and staged accordingly to git messages and configurations. The principle git practices were emphasized to remind the audience of when and how to commit changes to the previously specified local repository. Furthermore, Seiling prepared guests to make requests when merging work, added description templates for joint projects and generally taught the features of use for group collaboration.
This was followed by instructions on the key functionality of Git, such as the Git repository, Git commands, branches, and conflict resolution. For instance, the branches gave insight into how to leverage simultaneous work done separately from the overall code base. This is especially beneficial for feature development while also helping to streamline the process of reviewing changes before merging. Throughout this instruction, commands were given to switch branches and merge scripts in the terminal, which was operationalized with a quickly constructed example. Seiling also provided necessary information on managing repositories, including visuals of the basic workflow and linkage between local and remote repositories, either for individual or collaborative effort.
For those curious when to use which Git platform, Bukold jumped in to detail the major differences between Github, Gitlab and Git.
Terminal commands are used to perform actions with Git
Later, Seiling explored some advantageous elements of the GitLab platform, accessible free of charge to Weizenbaum researchers, by describing the repository graph, issue tracking and project management tools. To elaborate, the repository graph structures insight into how a participant makes a contribution or change by arranging branches to show merges or commits, particularly relevant for collaborative code projects. In case of software malfunctioning, the issue tracking feature allows one to see who is working on what branch for an update on the progress of the problem. Finally, Gitlab’s management tool was outlined for instances of assigning work, applying tags to notify when projects are finished and to open or close potential issues.
To close, Kostadinoski briefly summarized the basic elements of Git, along with its implications in data work, such as for software development and research. He simplified key terms and embraced questions in a Q&A. Seiling joined in, encouraging participants to “learn by doing” and stay connected with each other via Weizenbaum associated Github accounts for future internal coordination.
Throughout this workshop, participants were presented with various tasks and benefited from frequent recaps that highlighted key points, ensuring a solid understanding of the material. Attendees both online and in person freely asked questions and received support from instructors. Therefore the Methods Lab would like to give a huge thank you to LK Seiling, Sascha Kostadinoski and Quentin Bukold for their clear instruction on the foundations of Git and for facilitating such an engaging environment for all participants.
The Weizenbaum research groups “Digital Economy, Internet Ecosystem, and Internet Policy” (Jan Batzner) and “Data, Algorithmic Systems and Ethics” (Dr. Fatma Elsafoury), supported by Fraunhofer FOKUS and TU Berlin, with contributors Zeerak Talat and Flor Miriam Plaza del Arco, are excited to introduce the workshop “Social Science and Language Models – Methods and theory to responsible research on and with Language Technologies” taking place on April 3–4, 2025 at the Weizenbaum Institute. This hybrid event encourages interdisciplinary collaboration to promote ethically responsible research in the application of natural language technology. As methodology utilizing language models is increasingly applied to a variety of contexts from social science, health-care settings to computer software development, research suggests the growing need to monitor potential biased outcomes of its use. However, the absence of collaborative understanding between researchers of social science and those in Natural Language Processing (NLP), perpetuates discrimination as biases in the conception and measurement of socio-technical systems often go unrecognized.
Therefore we hope to engage a diverse group of researchers involved in the methodology of social or economic fields of discipline to address this prejudice in language technologies. Submissions of abstracts are encouraged to involve aspects of bias in the mitigation and measurement of NLP, as well as its implications in the social sciences.
In the first part of the event, participants discuss their experiences with networking strategies in a speed-dating format. Each conversation was documented by a member of the organizing team. Participants were rotated every few minutes to create different pairings. Each conversation was documented by a member of the organizing team. Participants highlighted the importance of networking within their own institutions, attending regularly organized events to formalize informal connections, pooling resources, and implementing cross-institutional research projects.
Melanie Althage (IZ D2MCM) guides participants through the new calendar system
In the second part of the event, colleagues from IZ D2MCM presented participants with a calendar system they developed. Its purpose is to consolidate events occurring at the network institutions into a single platform, making them accessible to all members. The system was then discussed in two groups. In one group, participants exchanged ideas on the design and admission criteria for events, considering aspects such as content, format, and location. In the other group, participants focused on facilitating the technical implementation, which operates through Git and enables network members to submit event metadata in a structured format.
The Methods Lab would like to thank the IZ D2MCM and all participants for their contributions to this successful event. Stay tuned for the next one!
After another successful run, the Methods Lab is excited to bring back the third annual Programming and Data Analysis with R workshop, led by Roland Toth (WI). This two-day event held at the Weizenbaum institute falls on Wednesday, March12th and Thursday, March 13th.
On the first day, one can expect a comprehensive introduction to the fundamentals of programming, essential data wrangling techniques and Markdown integration. Following this, the second day emphasizes data analysis and incorporates hands-on application of datasets, enabling attendees to independently explore a relevant research topic. Throughout both days, participants will be presented with conceptual knowledge, coding techniques and basic subtasks for a practical and immersive learning experience.
Join us in our first workshop of 2025 for an Introduction to Git, held on Thursday, February 6th. This event will be taking place at the Weizenbaum Institute and welcomes Weizenbaum Institute members to participate.
LK Seiling, an associate researcher, IT administrator Sascha Kostadinoski, and student assistant Quentin Bukold will be the primary instructors leading this event. Together they will guide participants through short theoretical segments, introducing fundamental Git commands and version control concepts. In addition to learning the operations of key Gitlab features, this workshop encourages guests to participate in quizzes and incorporates interactive exercises.
For further details, visit our program page. We hope to see you there!
Manage Cookie Consent
To provide the best experiences, we use cookies to store and/or access device information. Consenting to these technologies will allow us to process data such as browsing behavior or unique IDs on this site. Not consenting or withdrawing consent, may adversely affect certain features and functions.
Functional
Always active
The technical storage or access is strictly necessary for the legitimate purpose of enabling the use of a specific service explicitly requested by the subscriber or user, or for the sole purpose of carrying out the transmission of a communication over an electronic communications network.
Preferences
The technical storage or access is necessary for the legitimate purpose of storing preferences that are not requested by the subscriber or user.
Statistics
The technical storage or access that is used exclusively for statistical purposes.The technical storage or access that is used exclusively for anonymous statistical purposes. Without a subpoena, voluntary compliance on the part of your Internet Service Provider, or additional records from a third party, information stored or retrieved for this purpose alone cannot usually be used to identify you.
Marketing
The technical storage or access is required to create user profiles to send advertising, or to track the user on a website or across several websites for similar marketing purposes.