-
Notifications
You must be signed in to change notification settings - Fork 0
BEST: BERT Pre-training for Sign Language Recognition with Coupling Tokenization #19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Official Citation:
|
Reformatted:
|
Abstract:
|
c.f. #14 SignBERT+ figure SignBERT+ uses ONLY hand poses. They say: "We organize the pre-extracted 2D poses of both hands as the visual token sequence." Also different: SignBERT+ talks about joint, frame, and clip masking, and it sounds like BEST doesn't do different levels of masking, just the frame-level? |
Back to BEST:
|
Datasets:
|
Honestly, I think this one should not be merged until #14 is merged |
Interesting:
OK, what's that mean? |
|
|
"Pose Triplet Unit"? |
Oh this seems important, they're using a d-VAE. c.f. sign-language-processing#37
|
|
OK, in my own words now, real informally: So the thing to know about BEST is that they wanted to do BERT-style masked language modeling, but, you know, BERT assumes you've already got discrete, sematically meaningful tokens. |
OK, here's the transcript of the ChatGPT-assisted rewrite process: https://chatgpt.com/share/f42199e8-bd03-4ec3-9913-e5e832bb2885 My original summary was flawed in a number of ways.
Here's the first rewrite suggestions:
Here's the second:
|
In addition the PR had various suggestions, e.g. sign-language-processing#61 (comment) |
|
Merged |
https://ojs.aaai.org/index.php/AAAI/article/view/25470
dataset:
. Exclude wordy abstracts. (better BibTex extension to Zotero can exclude keys){}
in the bibtexPR:
git merge master
on branchWriting/style:
The text was updated successfully, but these errors were encountered: