-
Notifications
You must be signed in to change notification settings - Fork 24
Description
Hey,
I'm currently working on OMR with a focus on Turkish music notation, and I've been digging the oemer repo for a while now. I've also built a few proof-of-concept experiments based on it.
I noticed that apart from the common models like U-Net and SegNet, there are several custom ones in the repo—like clef.model, rests_above8.model, all_rests.model and sfn.model etc.. From what I gather, these are used to differentiate between various types of musical symbols.
Just wanted to ask—are you following the same logic for symbol differentiation in your implementation, or have you taken a different approach?
If it's the same, could you possibly share a bit more detail on how the training process works for those custom models—like how the datasets are extracted from ds2, structured, etc.?
Thanks indeed