Stylometry, or analysis of countable linguistic features of (literary) texts, is usually associated with authorship attribution or, more sensationally, with the sole purpose of prosecuting plagiarists. However, recent studies show that the same methods that help attribute an ancient Greek text to Plato or put a plagiarist behind bars for forging an alleged play by Shakespeare can be used in a much broader context of literary study. Patterns of stylometric similarity and difference provide new insights into relationships between different books by the same author; between books by different authors, between authors differing in terms of chronology or gender; between translations of the same author or group of authors; helping, in turn, to find new ways of looking at works that seem to have been studied from all possible perspectives.
The workshop, split into two 16-hour blocks, will try to address some of the following research questions: What is common in the language we use and what is related to cultural contexts and/or writer's individuality? What elements of style are affected by literary period, genre, topic? What is unconsciously incorporated by the author and reflects his/her education, gender, religious background, social or historical conditions? Last but not least, which features of a written text can betray the person who wrote it despite his/her aesthetic, social, or historical conditions?
The first and less advanced block, From Nothing to Networks, will acquaint the participants with 'stylo', the suite of stylometric tools produced by the workshop leaders, and with Gephi, a tool for network analysis.
The second block, From Networks to Nearest Shrunken Centroids, will introduce some more advanced problems of classification and validation, and more advanced methods of statistical analysis of texts.
Both blocks will focus on tools based in the R statistical programming environment with user-friendly interfaces, so no expert knowledge of R in particular, or of programming in general is required for Block 1. The texts used for the workshops will be provided by the instructors and the participants are encouraged to bring their own; if necessary, the participants' individual corpora will be expanded as needed and as available (online or elsewhere). The texts will be literary, multilingual, and include both originals and translations. Block 2 will be building on the skills acquired in Block 1, but a catching-up session is planned for participants with some R and Gephi knowledge.
- Important dates
- XML-TEI encoding, structuring and rendering
- Compilation, Annotation and Analysis of Written Text Corpora
- Comparing Corpora
- Digital Editions and Editorial Theory
- Searching Linguistic Patterns in Large Text Corpora for Digital Humanities Research
- Lexicometric text analysis using CLARIN-D Webservices and R
- Spoken Language and Multimodal Corpora
- Digital Lexica, Terminological Databases and Encyclopaedias
- Exploring art and technology within contemporary network culture
- From Text to Map: Modeling Historical Humanties Data in Mapping Environments
- Project Management
- Data management for the humanities
- Digital Research Infrastructures in the Humanities: How to Use, Build and Maintain Them
- Teasers (public)
- Lectures (public)
- Projects & Posters (public)
- T-Shirt 2016
- Scientific Committee
- Refund policy
- Child Care