Psycholinguistically Motivated Tree-Adjoinging Grammar (PLTAG) Parser: This implementation contains a fully incremental PLTAG parser, with incremental semantic role labeling capability and discriminative reranking. The parser is described in Demberg et al. (2013) and in subsequent papers. Please also try the Demo of the PLTAG Parser.

WebExp: A software package for conducting experiments over the world-wide web. Web-based experimentation gives access to a large and varied set of potential subjects, and experiments can be administered without the overheads of lab setups, attendance schedules, and so on. WebExp is written in Java, and uses XML as the description language for defining experiments and storing results. The software is described in Keller et al. (2009).


PASCAL07 Center-click Annotation Dataset: This dataset provides center-click annotations collected on Amazon Mechanical Turk for all 20 classes of the whole trainval set of PASCAL VOC 2007. Each image is annotated by two different annotators for each class in the image. This results in 14,612 clicks in total for the 5,011 trainval images. We also provide the localizations produced by our center-click object localization approach. The approach and the dataset are described in Papadopoulos et al. (2017).

Verb Senses in Images (VerSe) Dataset: 3,518 images, each annotated with one of 90 verbs, and with the OntoNotes sense realized for a given verb in the image. The images are taken from two existing multimodal datasets (COCO and TUHOI). The dataset is described in Gella et al. (2016).

Pascal Objects Eye Tracking (POET) Dataset: 6,270 images from ten Pascal VOC 2012 objects classes (cat, dog, bicycle, motorbike, boat, aeroplane, horse, cow, sofa, diningtable). Each image is annotated with the eye movement record of five participants, whose task was to identify which object class was present in the image. The dataset is described in Papadopoulos et al. (2014).

Visual and Linguistic Treebank: 2,424 images with human-generated image descriptions; 341 of these images are also annotated with object boundaries and Visual Dependency Representations. The dataset is described in Elliott and Keller (2013).

Object Naming Dataset: 100 images with eye-tracking data from 24 participants performing an object naming task. The data includes manually annotated object boundaries and object labels produced by participants. The dataset is described in Clarke et al. (2013).

Task Classification Dataset: Eye-movement dataset containing 1,756 unique trials across the three tasks: visual search, image description, and object naming. For each trial, the following standard features are extracted: (a) number of fixations, (b) mean fixation duration, (c) mean saccade amplitude, and (d) percent of image covered by fixations assuming a 18 circle around the fixation position, proportion of dwell time on (e) faces, (f) bodies, and (g) objects. A set of 15 additional features is also provided. The dataset is described in Coco and Keller (2014).

Scan Pattern Dataset: An image description dataset, which contains the eye-movement data of 24 participants describing 24 visual scenes. It includes scan patterns, transcribed sentences, and pairwise similarity scores for scan patterns and sentences. The dataset is described in Coco and Keller (2012).

Padó Plausiblity Dataset: Plausibility judgments for 207 verbs, with two arguments each, annoated with PropBank and FrameNet 1.2 semantic roles. Likert-scale judgments of plausibility were obtained in a web-based experiment from 100 participants. The dataset is described in Padó et al. (2006).

Bigram Plausiblity Dataset: Plausibility judgments for seen and unseen adjective-noun, noun-noun, and verb-object bigrams (90 items each). Magnitude estimation judgments of plausibility were obtained in a web-based experiment from 27 to 40 participants per item. The dataset is described in Keller and Lapata (2003).

Lab Facilities

Joint Eyetracking Lab: This lab is designed for studying a range of human cognitive processes, including reading, speech, dialog, and visual processing. The setup in the lab is unique in that it contains two state of the art head-mounted eyetrackers, controlled by customized software that makes it possible to study how participants interact when they solve collaborative tasks. Of course the lab can also be used for single-participant eyetracking studies.

Perception Lab: This lab suite is designed for the experimental study of human cognition. It can also be used for experiments that investigate how humans interact with artificial cognitive systems. It includes specialized soft- and hardware for the accurate measurement of reaction times, and for the presentation, recording, and analysis of speech data.