CogVideoX-Fun-V1.1-Reward-LoRAs是通过奖励反向传播技术训练Lora,以优化CogVideoX-Fun-V1.1生成的视频,使其更好地与人类偏好保持一致。
模型名 | 基础模型 | 奖励模式 | 下载地址 | 描述 |
---|---|---|---|---|
CogVideoX-Fun-V1.1-5b-InP-HPS2.1.safetensors | CogVideoX-Fun-V1.1-5b | HPS v2.1 | Link | Official HPS v2.1 reward LoRA (rank=128 and network_alpha=64 ) for CogVideoX-Fun-V1.1-5b-InP. It is trained with a batch size of 8 for 1,500 steps. |
CogVideoX-Fun-V1.1-2b-InP-HPS2.1.safetensors | CogVideoX-Fun-V1.1-2b | HPS v2.1 | Link | Official HPS v2.1 reward LoRA (rank=128 and network_alpha=64 ) for CogVideoX-Fun-V1.1-2b-InP. It is trained with a batch size of 8 for 3,000 steps. |
CogVideoX-Fun-V1.1-5b-InP-MPS.safetensors | CogVideoX-Fun-V1.1-5b | MPS | Link | Official MPS reward LoRA (rank=128 and network_alpha=64 ) for CogVideoX-Fun-V1.1-5b-InP. It is trained with a batch size of 8 for 5,500 steps. |
CogVideoX-Fun-V1.1-2b-InP-MPS.safetensors | CogVideoX-Fun-V1.1-2b | MPS | Link | Official MPS reward LoRA (rank=128 and network_alpha=64 ) for CogVideoX-Fun-V1.1-2b-InP. It is trained with a batch size of 8 for 16,000 steps. |
评论0