Skip to content

Commit a433a17

Browse files
authored
Dataset/LSA-T (#88)
* LSA-T JSON and citation * CDL: adding a comment * CDL: requested changes for PR
1 parent 9d3c5ab commit a433a17

File tree

3 files changed

+33
-0
lines changed

3 files changed

+33
-0
lines changed

src/datasets/LSA-T.json

+15
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,15 @@
1+
{
2+
"pub": {
3+
"name": "LSA-T",
4+
"year": 2022,
5+
"publication": "dataset:dal2022lsa",
6+
"url": "https://midusi.github.io/LSA-T/"
7+
},
8+
"features": ["video:RGB","text:es","pose:AlphaPose"],
9+
"language": "Argentina",
10+
"#items": null,
11+
"#samples": "14,880 sentences",
12+
"#signers": 103,
13+
"license": "MIT",
14+
"licenseUrl": "https://github.com/midusi/LSA-T/blob/main/LICENCE"
15+
}

src/index.md

+4
Original file line numberDiff line numberDiff line change
@@ -1028,6 +1028,10 @@ Research papers which do not necessarily contribute new theory or architectures
10281028

10291029
@dataset:joshi-etal-2023-isltranslate introduce ISLTranslate, a large translation dataset for Indian Sign Language based on publicly available educational videos intended for hard-of-hearing children, which happen to contain both Indian Sign Language and English audio voiceover conveying the same content. They use a speech-to-text model to transcribe the audio content, which they later manually corrected with the help of accompanying books also containing the same content. They also use MediaPipe to extract pose features, and have a certified ISL signer validate a small portion of the sign-text pairs. They provide a baseline based on the architecture proposed in @camgoz2020sign, and provide code.
10301030

1031+
<!-- TODO: LSA-T aka dataset:dal2022lsa, they use AlphaPose "with the Halpe full-body keypoints format", a visualizer tool, and a baseline SLT model. Especially might be good to mention FiftyOne https://docs.voxel51.com/, "which
1032+
provides useful features such as allowing to filter samples by label, video, playlist,
1033+
or by the confidence score of the signer inference." -->
1034+
10311035
###### Bilingual dictionaries {-}
10321036
for signed language [@dataset:mesch2012meaning;@fenlon2015building;@crasborn2016ngt;@dataset:gutierrez2016lse] map a spoken language word or short phrase to a signed language video.
10331037
One notable dictionary, SpreadTheSign\footnote{\url{https://www.spreadthesign.com/}} is a parallel dictionary containing around 25,000 words with up to 42 different spoken-signed language pairs and more than 600,000 videos in total. Unfortunately, while dictionaries may help create lexical rules between languages, they do not demonstrate the grammar or the usage of signs in context.

src/references.bib

+14
Original file line numberDiff line numberDiff line change
@@ -3443,3 +3443,17 @@ @misc{SiMAX2020SignLanguage
34433443
url = {https://cordis.europa.eu/project/id/778421},
34443444
urldate = {2024-06-18}
34453445
}
3446+
3447+
@inproceedings{dataset:dal2022lsa,
3448+
address = {Berlin, Heidelberg},
3449+
author = {Dal Bianco, Pedro and R\'{\i}os, Gast\'{o}n and Ronchetti, Franco and Quiroga, Facundo and Stanchi, Oscar and Hasperu\'{e}, Waldo and Rosete, Alejandro},
3450+
booktitle = {Advances in Artificial Intelligence – IBERAMIA 2022: 17th Ibero-American Conference on AI, Cartagena de Indias, Colombia, November 23–25, 2022, Proceedings},
3451+
doi = {10.1007/978-3-031-22419-5_25},
3452+
isbn = {978-3-031-22418-8},
3453+
numpages = {12},
3454+
pages = {293–304},
3455+
publisher = {Springer-Verlag},
3456+
title = {{LSA-T}: The First Continuous {Argentinian Sign Language} Dataset for Sign Language Translation},
3457+
url = {https://doi.org/10.1007/978-3-031-22419-5_25},
3458+
year = {2023}
3459+
}

0 commit comments

Comments
 (0)