
I really appreciate your research work. I have some questions regarding Table 1. It seems that UniTok needs to expand the codebook size or vocabulary size to 2^56 to achieve an rFID of 0.33. In this case, how can the GPT model be trained? Won’t the complexity be too high?