Kick off your journey into Automatic Text Recognition (ATR) with our introductory tutorial video. This is the first video of a tutorial series dedicated to extracting full text from scanned images.
This lesson uses word embeddings and clustering algorithms in Python to identify groups of similar documents in a corpus of approximately 9,000 academic abstracts. It will teach you the basics of dimensionality reduction for extracting structure from a large corpus and how to evaluate your results.
This lesson provides a beginner-friendly introduction to convolutional neural networks (CNNs) for image classification. The tutorial provides a conceptual understanding of how neural networks work by using Google’s Teachable Machine to train a model on paintings from the ArtUK database. This lesson also demonstrates how to use Javascript to embed the model in a live website.
This tutorial explores where and how to find, create, and collect images of textual material, a crucial initial step in any process using Automatic Text Recognition (ATR).
In this lesson, you will learn how to apply a Generative Pre-trained Transformer language model to a large-scale corpus so that you can locate broad themes and trends within written text.
This resource from the CLS INFRA project offers an introduction to several research areas and issues that are prominent withinComputational Literary Studies (CLS), including authorship attribution, literary history, literary genre, gender in literature, and canonicity/prestige, as well as to several key methodological concerns that are of importance when performing research in CLS.
This is the first of a two-part lesson introducing deep learning based computer vision methods for humanities research. Using a dataset of historical newspaper advertisements and the fastai Python library, the lesson walks through the pipeline of training a computer vision model to perform image classification.
This is the second of a two-part lesson introducing deep learning based computer vision methods for humanities research. This lesson digs deeper into the details of training a deep learning based computer vision model. It covers some challenges one may face due to the training data used and the importance of choosing an appropriate metric for your model. It presents some methods for evaluating the performance of a model.
Natural language processing is one of the most powerful concepts in modern linguistics and computer science, bridging the understanding of language from human to machine, and in turn programming machines so they can perform complex linguistic tasks on their own. This short video introduces learners to the key concepts of word embeddings and how they can be used in digital humanities projects.
In this lecture from the Austrian Centre for Digital Humanities and Cultural Heritage (ACDH-CH), Dirk Hovy gives an introduction to the method called embeddings, and showcases several applications of it. Hovy shows how they capture regional variation at an intra- and interlingual level, how they distinguish varieties and linguistic resources, and how they allow for the assessment of changing societal norms and associations.