In this lecture from the Austrian Centre for Digital Humanities and Cultural Heritage (ADCH-CH) entitled "Digitality and Music Editions". Thinking about the Roles of Editors, Stefanie Acquavella-Rauch discusses digitality as a method as well as a phenomenon and its role in music editions. In this talk, digitality is discussed as inhibiting a significant role in how researchers' roles change according to their personal engagement with the digital.
This video features Kathryn Sutherland, Professor of Bibliography and Textual Criticism at the University of Oxford, talking about the Jane Austen Fiction Manuscripts Project.
This video features speech technologist Henk van den Heuvel, linguist Silvia Calamai and data curator Louise Corti explaining how speech technology has reached the stage of being able to automatically recognise and retrieve speech in huge amounts of audio visual data.
This video features Paul Eggert, Loyola University Chicago, Kenneth M. Price, University of Nebraska, and Anne Baillot, Centre Marc Bloch, Berlin, talking about textual scholarship from analogue to digital.
This course will introduce you to the creation of digital scholarly editions, for manuscripts or printed texts, with the help of the TEI and other related technologies.
This video features Paul Eggert, Martin J. Svaglic Endowed Chair in Textual Studies, Department of English, Loyola University Chicago, talking about textual studies and the study of versions as a methodology to ask questions revealing the lifespan of a text.
This video features Dr. James Cummings, University of Oxford, Dr. Anne Baillot, Centre Marc Bloch, Dr. Marjorie Burghart, Centre National de la Recherche Scientifique CNRS, Prof Kenneth M. Price, University of Nebraska, and Prof Elena Pierazzo, Université Grenoble Alpes, interpreting what is textual scholarship and textual criticism.
This course is an introduction to the theories, practices, and methods of digitizing legacy dictionaries for research, preservation and online distribution. It focuses on a particular technique of modeling and describing lexical data using eXtensible Markup Language (XML) in accordance with the Guidelines of the Text Encoding Initiative, a de-facto standard for text encoding among humanities researchers.