Attention processor using custom rmsnorm kernel for Q/K normalization. NOTE: attn.norm_q and attn.norm_k HAVE weights (elementwise_affine=True).
Some results have been hidden because they may be inaccessible to you
Show inaccessible resultsSome results have been hidden because they may be inaccessible to you
Show inaccessible results