DepositThe Visual Page

All printed texts convey meaning through both linguistic and graphic signs, but existing tools for computational text analysis focus only on the linguistic content. The Visual Page will develop a prototype application to identify and analyze visual features in digitized Victorian books of poetry, such as margin space, line indentation, and typeface attributes. This will enable scholars to compare documents, identify distinctive or typical books, and track historical changes and influence over very large sets of digitized texts. Current research into such questions is limited by our human capacity to view and compare only a fairly small number of texts at one time. Thus our understanding of their historical significance is based on limited information. Computer analysis can point to significant patterns and trends over a much larger set of texts, which will ultimately transform our understanding of Victorian print culture and the humanities at large.

DepositEvolutionary Subject Tagging in the Humanities

Interdisciplinary research in the humanities requires indexing that represents multiple disciplinary perspectives. Most literature has been indexed using traditional models for subject analysis that are either too broad to be helpful or represent a single disciplinary perspective. We question whether traditional print models of subject analysis serve humanistic researchers’ needs in working with digital content. It is beyond the capacity of libraries to re-index this body of literature relying on human indexers. We need to develop scalable tools to both re-index extant bodies of literature and newly created literature. Web-scale searching, computational text analysis, and automated indexing each hold promise for addressing various aspects of the problem, but none seem to fully address the problem. This project will gather a group of scholars with expertise in the humanities, computational analysis of texts, and library and information science, to design an approach to the problem.

DepositGeoparsing history: Locating commodities in ten million pages of nineteenth-century sources

In the Trading Consequences project, historians, computational linguists, and computer scientists collaborated to develop a text mining system that extracts information from a vast amount of digitized published English-language sources from the “long nineteenth century” (1789 to 1914). The project focused on identifying relationships within the texts between commodities, geographical locations, and dates. The authors explain the methodology, uses, and the limitations of applying digital humanities techniques to historical research, and they argue that interdisciplinary approaches are critically important in addressing the technical challenges that arise. Collaborative teamwork of the kind described here has considerable potential to produce further advances in the large-scale analysis of historical documents.

MemberPatricia Murrieta-Flores

I’m the Director of the Digital Humanities Research Centre based at the Faculty of Humanities at the University of Chester. I’m also ERC Senior Researcher at the ‘Past in its Place Project’ (2014-2016) and Lecturer in Digital Humanities (from 2017). I’m part of the team of the HERA ‘Deepdead Project’ (2016-2019), a collaborator in the ‘Spatial Humanities Project’ at Lancaster University, and the European Cost Action ‘Reassembling the Republic of Letters’. My interest lies in the application of technologies for Humanities and her primary area of research is the Spatial Humanities and the investigation of different aspects of space, place and time using a range of technologies including GIS and Corpus Linguistic approaches. See some of my publications here: Patricia Murrieta-Flores in Academia.

DepositImage Analysis for Archival Discovery (Aida)

Images created in the digitization of primary materials contain a wealth of machine-processable information for data mining and large-scale analysis, and this information should be leveraged both to connect researchers with the resources they need and to augment interpretation of human culture, as a complement to and extension of text-based approaches. The proposed project, “Image Analysis for Archival Discovery” (Aida), applies image processing and machine learning techniques from computer science to digitized materials to facilitate and promote archival discovery. Beginning with the automatic detection of poetic content in historic newspapers, this project will develop image processing as a methodology for humanities research and analysis. In doing so, it will advance work on two fronts: 1) it will contribute to the reevaluation of newspaper verse in American literary history; 2) it will assess the application of image analysis as a method for discovery in archival collections.

DepositInstitute for High Performance Sound Technologies for Access and Scholarship (HiPSTAS)

We are applying for an Institutes for Advanced Technologies in the Digital Humanities grant from the NEH to support bringing together librarians and archivists, humanities scholars and students, and computer scientists and technologists invested in understanding and developing infrastructure for computational analysis on poetry, folklore, speeches, and storytelling sound files. The School of Information at the University of Texas at Austin and the Informatics Institute at the University of Illinois at Urbana-Champaign propose to host the High Performance Sound Technologies for Access and Scholarship (HiPSTAS) Institute to include meetings in May 2013 and in May 2014. In the interim year, scholars will work on scholarship in consultation with the HiPSTAS team. The second meeting is a symposium on the scholarship produced through the year as well as a meeting to propose recommendations for the development of tools for supporting advanced digital scholarly inquiry in spoken text sound.

MemberScott Kleinman

I am Professor of English at California State University, Northridge, where I have taught since August 1999. Prior to coming to Northridge, I taught at the University of Missouri, Columbia. I work on medieval language and literature from the Anglo-Saxon period to the fourteenth century with a special emphasis on Old English and early Middle English. My early work was on the history of the English language during the Old English period, especially the development of phonology and its dialects. More recently I have worked on regional and cultural diversity in historiographical and romance literature. I have a strong interest in Digital Humanities, particularly computational text analysis and digital editing. 

TopicCall for Papers: Digital Art History – Where Are We Now?

Special issue of Visual Resources In 2013, Visual Resources published a special issue devoted to Digital Art History. We recognize that since that date considerable activity has taken place in this area, which was then still in a phase of relative infancy. We feel that now is an opportune moment to assess what has […]

DepositLexomic Tools and Methods for Textual Analysis: Providing Deep Access to Digitized Texts

This project hybridizes traditional humanistic approaches to textual scholarship, such as source study and the analysis of style, with advanced computational and statistical comparative methods, allowing scholars “deep access” to digitized texts and textual corpora. Our multi-disciplinary collaboration enables us to discover patterns in (and between) texts previously invisible to traditional methods. Going forward, we will build on the success of our previous Digital Humanities Start-up Grant by further developing tools and documentation (in an open, on-line community) for applying advanced statistical methodologies to textual and literary problems. At the same time we will demonstrate the value of the approach by applying the tools and methods to texts from a variety of languages and time periods, including Old English, medieval Latin, and Modern English works from the twentieth-century Harlem Renaissance.