YouTube-ASL Clip Keypoint Dataset

Please use the following text to cite this item or export to a predefined format:
Železný, Tomáš; Hrúz, Marek; Straka, Jakub and Gueuwou, Shester, 2025, YouTube-ASL Clip Keypoint Dataset, DSpace at University of West Bohemia, http://hdl.handle.net/20.500.14592/98
Date issued
2025
Description
The YouTube-ASL Clip Keypoint Dataset is a curated collection of sentence-level American Sign Language (ASL) keypoint sequences derived from publicly available YouTube videos. Rather than providing raw video files, the dataset consists solely of JSON files containing frame-by-frame 2D keypoints extracted from segmented clips of individual signed sentences. Each frame has been processed using MediaPipe, which generates 208 2D keypoints representing body, face, hands, and pose landmarks. These keypoint sequences provide a compact, privacy-preserving representation of ASL visual-linguistic content, enabling research in sign language recognition, gesture analysis, and multimodal communication. The dataset consists of 390 547 json files zipped in 10 separate zip files for easier manipulation. Beside the keypoint files, we also provide the annotation json files.
This item isPublicly Available
and licensed under:
 Files in this item