I have run the interference from ASL signs on NGT videos, and i have to say the model has produced some surprising results.
The model recognized dutch signs and produced glosses from similair ASL signs.
Now I want to prepare dataset from Corpus NGT, signCollect (which consist 4K 60FPS quality) and Signbank.
- signCollect already has videos with gloss name, we have about 400 of them with left/mid/right
- Signbank
- Corpus NGT, I am thinking to segment the videos with annotated EAF files and then export the videos to seperate folders named by gloss
And then generate parque files from them, linked to metadata JSON file consisting of gloss ID, frame start/end, frames count and fps
Then i have to run the training script from asl competition.
@AmitMY is the preparation I have in my mind right?