'Digital Humanities books’ by M C Morgan, Flickr (CC BY-SA 2.0)
What is Digital Humanities?
While there is much discussion about the definition of Digital Humanities (DH), broadly defined, it is an area of scholarly activity at the crossroads of digital technologies and the humanities. Originally known as “humanities computing”, it is where digital technologies are used to transform research and teaching across the humanities.
Digital Humanities Methodologies and Tools
For the numerous research activities related to Digital Humanities see TaDiRAH.
These are more comprehensive listings of the digital research tools available.
The DiRT Directory aggregates information about digital research tools for scholarly use. A tool, service, and the most comprehensive and up-to-date collection registry of digital research tools for scholarly use.
Contains curated lists of tools and code for studying texts. Project is led and based at the University of Alberta.
Beginner’s tool box prepared by Miriam Posner, American Studies Association.
Curated by Alan Liu, English professor at the University of California, Santa Barbara.
More than simply tools, this site also contains tutorials on how to use them.
“Digital tools for discerning historians”.
These are a selection of some of the most useful tools. In the main they are free and open source; where they aren’t a free trial is usually available. They are organised under the following headings: Capture, Creation, Enrichment, Analysis and Dissemination.
Scribe offers an open source framework for setting up community transcription projects around handwritten or OCR-resistant texts. Scribe is particularly geared toward digital humanities, library, and citizen science projects seeking to extract highly structured, normalisable data from a set of digitised materials.
T-PEN (Transcription for Paleographical and Editorial Notation)
Web-based set of tools to allow collaborative transcription of manuscript pages in TEI-compliant XML. Users attach transcription data (new or uploaded) to the actual lines of the original manuscript in a simple, flexible interface.
Text recognition software.
A web service to manage the transformation of documents between a variety of formats. The majority of transformations use the Text Encoding Initiative format as a pivot format.
Pandoc can convert documents in reStructuredText, textile, HTML, or LaTeX formats to a variety of other formats including XHTML, PDF, EPUB, docx, odt, and more.
Turtle, the Terse RDF Triple Language, a concrete syntax for RDF. Developed by David Beckett and Tim Berners-Lee. Turtle does not rely on XML and is more readable and easier to edit manually than XML and SPARQL, the query language for RDF, uses a similar syntax to Turtle for expressing query patterns.
Open source, web-based app that integrates a powerful set of textual interpretation tools behind an intuitive interface. You can upload your texts and annotate them with styled text, video, images and weblinks.
An online environment that synchronises web based video with timeline based annotations.
Tool for cleaning messy data (e.g. fixing inconsistencies), transforming between different formats and exploring data. Previously Google Refine.
The TEI is a consortium that develops and maintains a standard for the representation of texts in digital form. Its chief deliverable is a set of Guidelines that specify encoding methods for machine-readable texts, chiefly in the humanities, social sciences and linguistics.
Digital scholarly editing tool with long term research data archive.
Microsoft XML notepad is an open-source XML editor.
Textual analysis tools
CATMA Computer Aided Textual Markup and Analysis.
A free, open source markup and analysis tool. Also generates basic visualisation options for texts and corpora. Interfaces with the Voyant toolset.
Program which can search Google’s text corpora for a single word or phrase in sources printed between 1500 and 2008.
Open source tool for comparing and collating multiple witnesses to a single textual work. You can upload your comparison sets to a free online workspace, Juxta Commons, where you can analyse your data privately or share visualisations of your work with anyone on the web.
TAPoRware is a set of text analysis tools that enable users to perform text analysis on HTML, XML and plain text files using documents form the users’ machine or on the Web.
Free text analysis app that allows you to analyse webpages, tweet streams and documents; explores the relationships between words in the text via an intuitive word cloud interface.
Voyant is a web-based reading and analysis environment, designed to facilitate reading and interpretive practices. Voyant Tools is an open-source project and the code is available through GitHub.
A text analysis environment that combines visualisation, information retrieval, sense making and natural language processing to make the contents of text navigable, accessible, and useful.
A Java-based package for statistical natural language processing, document classification, clustering, topic modelling, information extraction, and other machine learning applications to text.
Weka is a collection of machine learning algorithms for data mining tasks.
A free, open-source template for Microsoft® Excel® 2007 and 2010 that makes it easy to explore network graphs.
Chronos Timeline allows you to dynamically present historical data in a flexible online environment.
Open source, allows you to create timelines and maps in minutes from a spreadsheet.
Visualisation and exploration software for all kinds of graphs and networks; open source and free.
A platform for visualising and analysing networks of historical data.
Free, open source data curation and visualisation plugin for WordPress.
A free data storytelling app where you can create and share interactive charts and graphs, maps and live dashboards.
App which can be used to generate graphs and statics and share the data and visualisations.
A closed source, ArcGIS is a platform for building a complete geographic information system (GIS) that lets you easily create, edit, and analyse geographic knowledge. Free trial is available.
Mapping tool with animation.
Mapping tool which allows you to collect spatio-temporal data and import it. Can be used to composed “MapStories” combining datasets with other narrative elements such as images, text or video.
A plugin for Omeka, Neatline is a tool for the creation of interlinked timelines and maps as interpretive expressions of the literary or historical content of archival collections.
Free editable map of the world.
A community and infrastructure for Linked Open Geodata in the Humanities.
User-friendly, open source, geographic Information System; supports numerous vector, raster and database formats and functionalities.
Open source software developed by the Center for Geographic Analysis at Harvard. Allows you to build your own mapping portal and share it.
Open source content management system for supporting resources like blogs and web sites.
Omeka is web-publishing platform that allows anyone with an account to create or collaborate on a website to display collections and build digital exhibitions. A free basic plan is available for a hosted account; Omeka downloadable is also free.
Scalar is a free, open source authoring and publishing platform.
Free, easy-to-use web publishing platform originally designed around blogging that has now evolved with functionality as a robust content or learning management system, with many themes and plugins for extra functionality, e.g. DH Press: a digital humanities toolkit designed for non-technical users; CommentPress allows readers to comment paragraph-by-paragraph, line-by-line or block-by-block in the margins of a text.