Thanks for the great work! I was wondering if non-causal attention will be supported in the near future to accommodate VLMs. Currently, only causal attention is available. https://github.com/zhuzilin/ring-flash-attention/blob/786677930bce4f6022166899c88ce2c00c814ee2/ring_flash_attn/adapters/hf_adapter.py#L142