상세 정보
추천
v1.0
SPO-SD-v1-5_4k-p_10ep_LoRA_webui

SPO-SD-v1-5_4k-p_10ep_LoRA_webui

359
5
190
#기본 모델

Aesthetic Post-Training Diffusion Models from Generic Preferences with Step-by-step Preference

Arxiv Paper

Github Code

Project Page

Abstract

Generating visually appealing images is fundamental to modern text-to-image generation models. A potential solution to better aesthetics is direct preference optimization (DPO), which has been applied to diffusion models to improve general image quality including prompt alignment and aesthetics. Popular DPO methods propagate preference labels from clean image pairs to all the intermediate steps along the two generation trajectories. However, preference labels provided in existing datasets are blended with layout and aesthetic opinions, which would disagree with aesthetic preference. Even if aesthetic labels were provided (at substantial cost), it would be hard for the two-trajectory methods to capture nuanced visual differences at different steps.

To improve aesthetics economically, this paper uses existing generic preference data and introduces step-by-step preference optimization (SPO) that discards the propagation strategy and allows fine-grained image details to be assessed. Specifically, at each denoising step, we 1) sample a pool of candidates by denoising from a shared noise latent, 2) use a step-aware preference model to find a suitable win-lose pair to supervise the diffusion model, and 3) randomly select one from the pool to initialize the next denoising step. This strategy ensures that diffusion models focus on the subtle, fine-grained visual differences instead of layout aspect. We find that aesthetic can be significantly enhanced by accumulating these improved minor differences.

When fine-tuning Stable Diffusion v1.5 and SDXL, SPO yields significant improvements in aesthetics compared with existing DPO methods while not sacrificing image-text alignment compared with vanilla models. Moreover, SPO converges much faster than DPO methods due to the step-by-step alignment of fine-grained visual details. Code and model: https://rockeycoss.github.io/spo.github.io/

Model Description

This model is fine-tuned from runwayml/stable-diffusion-v1-5. It has been trained on 4,000 prompts for 10 epochs. This checkpoint is a LoRA checkpoint. For more information, please visit here

Citation

If you find our work useful, please consider giving us a star and citing our work.

@article{liang2024step,
  title={Aesthetic Post-Training Diffusion Models from Generic Preferences with Step-by-step Preference Optimization},
  author={Liang, Zhanhao and Yuan, Yuhui and Gu, Shuyang and Chen, Bohan and Hang, Tiankai and Cheng, Mingxi and Li, Ji and Zheng, Liang},
  journal={arXiv preprint arXiv:2406.04314},
  year={2024}
}

번역문 보기

평점 및 리뷰

-- /5
0 개의 평점

충분한 평가나 댓글을 받지 못했습니다.

no-data
데이터 없음
R
모델과 대화하기
공고
2024-06-20
모델 게시
2024-06-20
모델 정보 업데이트
모델 상세정보
유형
LORA
게시 날짜
2024-06-20
기본 모델
SD 1.5
허가 범위
모델 출처: civitai

1. 재게시된 모델의 권리는 원 제작자에게 있습니다.

2. 모델 원작자가 모델을 인증받으려면 공식 채널을 통해 SeaArt.AI 직원에게 문의하세요. 저희는 모든 창작자의 권리를 보호하기 위해 노력합니다. 인증하러 이동

창작 허가 범위
온라인 생방송
혼합 진행
다운로드 허용
상업적 허가 범위
생성된 이미지를 판매하거나 상업적 목적으로 사용 가능
모델의 재판매 또는 융합 후 판매 허용
QR Code
SeaArt 앱 다운로드
모바일에서 AI 창작 여정을 계속하세요