-
Notifications
You must be signed in to change notification settings - Fork 71
Description
Thank you for releasing such an excellent project. I found WAM/VideoSeal and the provided model variants (PixelSeal, ChunkySeal, etc.) very interesting, and I’d like to retrain models for robustness against additional distortions.
I have a question about retraining the models referenced in videoseal/cards/*.yaml:
1. Are the cards intended as inference model cards only, or are they also meant to be full training recipes?
• Some cards (e.g., videoseal_0.0.yaml) look like they contain full training hyperparameters, but others (e.g., pixelseal.yaml, chunkyseal.yaml) seem to only specify architecture + a few key args (nbits, img_size_proc, attenuation, etc.).
• Is the intended way to retrain these variants to start from the docs/training.md commands and substitute the architecture/args from each card?
2. PixelSeal training: In the PixelSeal paper, the training pipeline looks more complex (multi-stage / specific loss).
• Does this repository include the full PixelSeal training pipeline/configs/scripts used to produce the released PixelSeal checkpoint?