Thanks to visit codestin.com
Credit goes to github.com

Skip to content

littlewhitesea/training-free-methods

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 
Β 
Β 

Repository files navigation

training-free-methods

This repository collects recent training-free algorithms relevant to visual generation and manipulation, with a focus on methods that can be run on a single GPU (ideally with ≀24GB memory). Hope it could be beneficial for students and researchers with limited computational resources.

If you find any missed resources or errors, feel free to open an issue or make a pull request.

πŸ“š Table of Contents

Image Generation

Luminark: Training-free, Probabilistically-Certified Watermarking for General Vision Generative Models.
Jiayi Xu, Zhang Zhang, Yuanrui Zhang, Ruitao Chen, Yixian Xu, Tianyu He, Di He.
arxiv 2026. [PDF]

Unraveling MMDiT Blocks: Training-free Analysis and Enhancement of Text-conditioned Diffusion.
Binglei Li, Mengping Yang, Zhiyu Tan, Junping Zhang, Hao Li.
arxiv 2026. [PDF]

ASemConsist: Adaptive Semantic Feature Control for Training-Free Identity-Consistent Generation.
Shin seong Kim, Minjung Shin, Hyunin Cho, Youngjung Uh.
arxiv 2025. [PDF] [Project] [Code]

AnyMS: Bottom-up Attention Decoupling for Layout-guided and Training-free Multi-subject Customization.
Binhe Yu, Zhen Wang, Kexin Li, Yuqian Yuan, Wenqiao Zhang, Long Chen, Juncheng Li, Jun Xiao, Yueting Zhuang.
arxiv 2025. [PDF]

TextGuider: Training-Free Guidance for Text Rendering via Attention Alignment.
Kanghyun Baek, Sangyub Lee, Jin Young Choi, Jaewoo Song, Daemin Park, Jooyoung Choi, Chaehun Shin, Bohyung Han, Sungroh Yoon.
arxiv 2025. [PDF]

SJD++: Improved Speculative Jacobi Decoding for Training-free Acceleration of Discrete Auto-regressive Text-to-Image Generation.
Yao Teng, Zhihuan Jiang, Han Shi, Xian Liu, Xuefei Ning, Guohao Dai, Yu Wang, Zhenguo Li, Xihui Liu.
arxiv 2025. [PDF]

Training-Free Generation of Diverse and High-Fidelity Images via Prompt Semantic Space Optimization.
Debin Meng, Chen Jin, Zheng Gao, Yanran Li, Ioannis Patras, Georgios Tzimiropoulos.
arxiv 2025. [PDF]

A Training-Free Approach for Multi-ID Customization via Attention Adjustment and Spatial Control.
Jiawei Lin, Guanlong Jiao, Jianjin Xu.
arxiv 2025. [PDF]

Infinite-Story: A Training-Free Consistent Text-to-Image Generation.
Jihun Park, Kyoungmin Lee, Jongmin Gim, Hyeonseo Jo, Minseok Oh, Wonhyeok Choi, Kyumin Hwang, Jaeyeul Kim, Minwoo Choi, Sunghoon Im.
AAAI 2026. [PDF]

TAUE: Training-free Noise Transplant and Cultivation Diffusion Model.
Daichi Nagai, Ryugo Morita, Shunsuke Kitada, Hitoshi Iyatomi.
arxiv 2025. [PDF] [Project] [Code]

ETC: training-free diffusion models acceleration with Error-aware Trend Consistency.
Jiajian Xie, Hubery Yin, Chen Li, Zhou Zhao, Shengyu Zhang.
arxiv 2025. [PDF]

GeoDiffusion: A Training-Free Framework for Accurate 3D Geometric Conditioning in Image Generation.
Phillip Mueller, Talip Uenlue, Sebastian Schmidt, Marcel Kollovieh, Jiajie Fan, Stephan Guennemann, Lars Mikelsons.
ICCV 2025. [PDF] [Code]

Stroke2Sketch: Harnessing Stroke Attributes for Training-Free Sketch Generation.
Rui Yang, Huining Li, Yiyi Long, Xiaojun Wu, Shengfeng He.
ICCV 2025. [PDF] [Code]

MosaicDiff: Training-free Structural Pruning for Diffusion Model Acceleration Reflecting Pretraining Dynamics.
Bowei Guo, Shengkun Tang, Cong Zeng, Zhiqiang Shen.
ICCV 2025. [PDF] [Code]

PEO: Training-Free Aesthetic Quality Enhancement in Pre-Trained Text-to-Image Diffusion Models with Prompt Embedding Optimization.
Hovhannes Margaryan, Bo Wan, Tinne Tuytelaars.
arxiv 2025. [PDF] [Code]

Stitch: Training-Free Position Control in Multimodal Diffusion Transformers.
Jessica Bader, Mateusz Pach, Maria A. Bravo, Serge Belongie, Zeynep Akata.
arxiv 2025. [PDF] [Code]

SemanticControl: A Training-Free Approach for Handling Loosely Aligned Visual Conditions in ControlNet.
Woosung Joung, Daewon Chae, Jinkyu Kim.
BMVC 2025. [PDF] [Project] [Code]

Training-Free Synthetic Data Generation with Dual IP-Adapter Guidance.
Luc Boudier, Loris Manganelli, Eleftherios Tsonis, Nicolas Dufour, Vicky Kalogeiton.
BMVC 2025. [PDF] [Project]

InfoScale: Unleashing Training-free Variable-scaled Image Generation via Effective Utilization of Information.
Guohui Zhang, Jiangtong Tan, Linjiang Huang, Zhonghang Yuan, Naishan Zheng, Jie Huang, Feng Zhao.
arxiv 2025. [PDF] [Code]

CountLoop: Training-Free High-Instance Image Generation via Iterative Agent Guidance.
Anindya Mondal, Ayan Banerjee, Sauradip Nag, Josep LladΓ³s, Xiatian Zhu, Anjan Dutta.
arxiv 2025. [PDF] [Project] [Code]

HiCache: Training-free Acceleration of Diffusion Models via Hermite Polynomial-based Feature Caching.
Liang Feng, Shikang Zheng, Jiacheng Liu, Yuqi Lin, Qinming Zhou, Peiliang Cai, Xinyu Wang, Junjie Chen, Chang Zou, Yue Ma, Linfeng Zhang.
arxiv 2025. [PDF]

S^2-Guidance: Stochastic Self Guidance for Training-Free Enhancement of Diffusion Models.
Chubin Chen, Jiashu Zhu, Xiaokun Feng, Nisha Huang, Meiqi Wu, Fangyuan Mao, Jiahong Wu, Xiangxiang Chu, Xiu Li.
arxiv 2025. [PDF] [Project] [Code]

CountCluster: Training-Free Object Quantity Guidance with Cross-Attention Map Clustering for Text-to-Image Generation.
Joohyeon Lee, Jin-Seop Lee, Jee-Hyong Lee.
arxiv 2025. [PDF] [Code]

Subject or Style: Adaptive and Training-Free Mixture of LoRAs.
Jia-Chen Zhang, Yu-Jie Xiong.
arxiv 2025. [PDF] [Code]

FreeLoRA: Enabling Training-Free LoRA Fusion for Autoregressive Multi-Subject Personalization.
Peng Zheng, Ye Wang, Rui Ma, Zuxuan Wu.
arxiv 2025. [PDF]

Story2Board: A Training-Free Approach for Expressive Storyboard Generation.
David Dinkevich, Matan Levy, Omri Avrahami, Dvir Samuel, Dani Lischinski.
arxiv 2025. [PDF] [Project] [Code]

LaRender: Training-Free Occlusion Control in Image Generation via Latent Rendering.
Xiaohang Zhan, Dingming Liu.
ICCV 2025. [PDF] [Project]

T2I-Copilot: A Training-Free Multi-Agent Text-to-Image System for Enhanced Prompt Interpretation and Interactive Generation.
Chieh-Yun Chen, Min Shi, Gong Zhang, Humphrey Shi.
ICCV 2025. [PDF] [Code]

FreeCus: Free Lunch Subject-driven Customization in Diffusion Transformers.
Yanbing Zhang, Zhe Wang, Qin Zhou, Mengping Yang.
ICCV 2025. [PDF] [Code]

Normalized Attention Guidance: Universal Negative Guidance for Diffusion Models.
Dar-Yen Chen, Hmrishav Bandyopadhyay, Kai Zou, Yi-Zhe Song.
arxiv 2025. [PDF] [Project] [Code]

A Training-Free Style-Personalization via Scale-wise Autoregressive Model.
Kyoungmin Lee, Jihun Park, Jongmin Gim, Wonhyeok Choi, Kyumin Hwang, Jaeyeul Kim, Sunghoon Im.
arxiv 2025. [PDF]

RichControl: Structure- and Appearance-Rich Training-Free Spatial Control for Text-to-Image Generation.
Liheng Zhang, Lexi Pang, Hang Ye, Xiaoxuan Ma, Yizhou Wang.
arxiv 2025. [PDF] [Project] [Code]

FreeMorph: Tuning-Free Generalized Image Morphing with Diffusion Model.
Yukang Cao, Chenyang Si, Jinghao Wang, Ziwei Liu.
ICCV 2025. [PDF] [Project] [Code]

HiWave: Training-Free High-Resolution Image Generation via Wavelet-Based Diffusion Sampling.
Tobias Vontobel, Seyedmorteza Sadat, Farnood Salehi, Romann M. Weber.
arxiv 2025. [PDF]

Control and Realism: Best of Both Worlds in Layout-to-Image without Training.
Bonan Li, Yinhan Hu, Songhua Liu, Xinchao Wang.
ICML 2025. [PDF]

LoRAShop: Training-Free Multi-Concept Image Generation and Editing with Rectified Flow Transformers.
Yusuf Dalva, Hidir Yesiltepe, Pinar Yanardag.
arxiv 2025. [PDF] [Project]

MultLFG: Training-free Multi-LoRA composition using Frequency-domain Guidance.
Aniket Roy, Maitreya Suin, Ketul Shah, Rama Chellappa.
arxiv 2025. [PDF]

ConsiStyle: Style Diversity in Training-Free Consistent T2I Generation.
Yohai Mazuz, Janna Bruner, Lior Wolf.
arxiv 2025. [PDF] [Project]

ISAC: Training-Free Instance-to-Semantic Attention Control for Improving Multi-Instance Generation.
Sanghyun Jo, Wooyeol Lee, Ziseok Lee, Kyungsu Kim.
arxiv 2025. [PDF]

ExpertGen: Training-Free Expert Guidance for Controllable Text-to-Face Generation.
Liang Shi, Yun Fu.
arxiv 2025. [PDF]

Training-Free Watermarking for Autoregressive Image Generation.
Yu Tong, Zihao Pan, Shuai Yang, Kaiyang Zhou.
arxiv 2025. [PDF] [Code]

Grouping First, Attending Smartly: Training-Free Acceleration for Diffusion Transformers.
Sucheng Ren, Qihang Yu, Ju He, Alan Yuille, Liang-Chieh Chen.
arxiv 2025. [PDF] [Project] [Code]

CountDiffusion: Text-to-Image Synthesis with Training-Free Counting-Guidance Diffusion.
Yanyu Li, Pencheng Wan, Liang Han, Yaowei Wang, Liqiang Nie, Min Zhang.
arxiv 2025. [PDF]

SphereDiff: Tuning-free Omnidirectional Panoramic Image and Video Generation via Spherical Latent Representation.
Minho Park, Taewoong Kang, Jooyeol Yun, Sungwon Hwang, Jaegul Choo.
arxiv 2025. [PDF] [Project] [Code]

FreeGraftor: Training-Free Cross-Image Feature Grafting for Subject-Driven Text-to-Image Generation.
Zebin Yao, Lei Ren, Huixing Jiang, Chen Wei, Xiaojie Wang, Ruifan Li, Fangxiang Feng.
arxiv 2025. [PDF] [Code]

Storybooth: Training-free Multi-Subject Consistency for Improved Visual Storytelling.
Jaskirat Singh, Junshen Kevin Chen, Jonas Kohler, Michael Cohen.
ICLR 2025. [PDF]

A Training-Free Style-aligned Image Generation with Scale-wise Autoregressive Model.
Jihun Park, Jongmin Gim, Kyoungmin Lee, Minseok Oh, Minwoo Choi, Jaeyeul Kim, Woo Chool Park, Sunghoon Im.
arxiv 2025. [PDF]

HiFlow: Training-free High-Resolution Image Generation with Flow-Aligned Guidance.
Jiazi Bu, Pengyang Ling, Yujie Zhou, Pan Zhang, Tong Wu, Xiaoyi Dong, Yuhang Zang, Yuhang Cao, Dahua Lin, Jiaqi Wang.
arxiv 2025. [PDF] [Code]

Training-free Dense-Aligned Diffusion Guidance for Modular Conditional Image Synthesis.
Zixuan Wang, Duo Peng, Feng Chen, Yuwei Yang, Yinjie Lei.
arxiv 2025. [PDF] [Code]

Spatial Transport Optimization by Repositioning Attention Map for Training-Free Text-to-Image Synthesis.
Woojung Han, Yeonkyung Lee, Chanyoung Kim, Kwanghyun Park, Seong Jae Hwang.
CVPR 2025. [PDF] [Project] [Code]

TF-TI2I: Training-Free Text-and-Image-to-Image Generation via Multi-Modal Implicit-Context Learning in Text-to-Image Models.
Teng-Fang Hsiao, Bo-Kai Ruan, Yi-Lun Wu, Tzu-Ling Lin, Hong-Han Shuai.
arxiv 2025. [PDF] [Project] [Code]

Personalize Anything for Free with Diffusion Transformer.
Haoran Feng, Zehuan Huang, Lin Li, Hairong Lv, Lu Sheng.
arxiv 2025. [PDF] [Project] [Code]

Localized Concept Erasure for Text-to-Image Diffusion Models Using Training-Free Gated Low-Rank Adaptation.
Byung Hyun Lee, Sungjin Lim, Se Young Chun.
CVPR 2025. [PDF]

EditID: Training-Free Editable ID Customization for Text-to-Image Generation.
Guandong Li, Zhaobin Chu.
arxiv 2025. [PDF]

ToLo: A Two-Stage, Training-Free Layout-To-Image Generation Framework For High-Overlap Layouts.
Linhao Huang, Jing Yu.
arxiv 2025. [PDF] [Code]

FlipConcept: Tuning-Free Multi-Concept Personalization for Text-to-Image Generation.
Young Beom Woo, Sun Eung Kim.
arxiv 2025. [PDF]

MagicGeo: Training-Free Text-Guided Geometric Diagram Generation.
Junxiao Wang, Ting Zhang, Heng Yu, Jingdong Wang, Hua Huang.
arxiv 2025. [PDF]

LoRA-X: Bridging Foundation Models with Training-Free Cross-Model Adaptation.
Farzad Farhadzadeh, Debasmit Das, Shubhankar Borse, Fatih Porikli.
ICLR 2025. [PDF]

Event-Customized Image Generation.
Zhen Wang, Yilei Jiang, Dong Zheng, Jun Xiao, Long Chen.
arxiv 2024. [PDF]

AP-LDM: Attentive and Progressive Latent Diffusion Model for Training-Free High-Resolution Image Generation.
Boyuan Cao, Jiaxin Ye, Yujie Wei, Hongming Shan.
arxiv 2024. [PDF] [Code]

Training-free Color-Style Disentanglement for Constrained Text-to-Image Synthesis.
Aishwarya Agarwal, Srikrishna Karanam, Balaji Vasan Srinivasan.
arxiv 2024. [PDF]

Smoothed Energy Guidance: Guiding Diffusion Models with Reduced Energy Curvature of Attention.
Susung Hong.
NeurIPS 2024. [PDF] [Code]

MegaFusion: Extend Diffusion Models towards Higher-resolution Image Generation without Further Tuning.
Haoning Wu, Shaocheng Shen, Qiang Hu, Xiaoyun Zhang, Ya Zhang, Yanfeng Wang.
arxiv 2024. [PDF] [Project] [Code]

DiffuseHigh: Training-free Progressive High-Resolution Image Synthesis through Structure Guidance.
Younghyun Kim, Geunmin Hwang, Junyu Zhang, Eunbyung Park.
arxiv 2024. [PDF] [Project] [Code]

TraDiffusion: Trajectory-Based Training-Free Image Generation.
Mingrui Wu, Oucheng Huang, Jiayi Ji, Jiale Li, Xinyue Cai, Huafeng Kuang, Jianzhuang Liu, Xiaoshuai Sun, Rongrong Ji.
arxiv 2024. [PDF] [Code]

MagicFace: Training-free Universal-Style Human Image Customized Synthesis.
Yibin Wang, Weizhong Zhang, Cheng Jin.
arxiv 2024. [PDF] [Project] [Code]

AccDiffusion: An Accurate Method for Higher-Resolution Image Generation.
Zhihang Lin, Mingbao Lin, Meng Zhao, Rongrong Ji.
ECCV 2024. [PDF] [Project] [Code]

ResMaster: Mastering High-Resolution Image Generation via Structural and Fine-Grained Guidance.
Shuwei Shi, Wenbo Li, Yuechen Zhang, Jingwen He, Biao Gong, Yinqiang Zheng.
arXiv 2024. [PDF] [Project] [Code]

Coherent Zero-Shot Visual Instruction Generation.
Quynh Phung, Songwei Ge, Jia-Bin Huang.
arXiv 2024. [PDF] [Project]

FreeCustom: Tuning-Free Customized Image Generation for Multi-Concept Composition.
Ganggui Ding, Canyu Zhao, Wen Wang, Zhen Yang, Zide Liu, Hao Chen, Chunhua Shen.
CVPR 2024. [PDF] [Code]

Training-free Subject-Enhanced Attention Guidance for Compositional Text-to-image Generation.
Shengyuan Liu, Bo Wang, Ye Ma, Te Yang, Xipeng Cao, Quan Chen, Han Li, Di Dong, Peng Jiang.
arXiv 2024. [PDF]

DemoFusion: Democratising High-Resolution Image Generation With No $$$.
Ruoyi Du, Dongliang Chang, Timothy Hospedales, Yi-Zhe Song, Zhanyu Ma.
CVPR 2024. [PDF] [Project] [Code]

HiDiffusion: Unlocking High-Resolution Creativity and Efficiency in Low-Resolution Trained Diffusion Models.
Shen Zhang, Zhaowei Chen, Zhenyu Zhao, Zhenyuan Chen, Yao Tang, Yuhao Chen, Wengang Cao, Jiajun Liang.
ECCV 2024. [PDF] [Project] [Code]

Training-Free Consistent Text-to-Image Generation.
Yoad Tewel, Omri Kaduri, Rinon Gal, Yoni Kasten, Lior Wolf, Gal Chechik, Yuval Atzmon.
SIGGRAPH 2024. [PDF] [Project]

FouriScale: A Frequency Perspective on Training-Free High-Resolution Image Synthesis.
Linjiang Huang, Rongyao Fang, Aiping Zhang, Guanglu Song, Si Liu, Yu Liu, Hongsheng Li.
ECCV 2024. [PDF] [Code]

Training-Free Layout Control with Cross-Attention Guidance.
Minghao Chen, Iro Laina, Andrea Vedaldi.
WACV 2024. [PDF] [Project] [Code]

BoxDiff: Text-to-Image Synthesis with Training-Free Box-Constrained Diffusion.
Jinheng Xie, Yuexiang Li, Yawen Huang, Haozhe Liu, Wentian Zhang, Yefeng Zheng, Mike Zheng Shou.
ICCV 2023. [PDF] [Code]

πŸ” Return to Top

Image Manipulation

TalkPhoto: A Versatile Training-Free Conversational Assistant for Intelligent Image Editing.
Yujie Hu, Zecheng Tang, Xu Jiang, Weiqi Li, Jian Zhang.
arxiv 2026. [PDF]

FreeText: Training-Free Text Rendering in Diffusion Transformers via Attention Localization and Spectral Glyph Injection.
Ruiqiang Zhang, Hengyi Wang, Chang Liu, Guanjie Wang, Zehua Ma, Weiming Zhang.
arxiv 2026. [PDF]

FreeInpaint: Tuning-free Prompt Alignment and Visual Rationality Enhancement in Image Inpainting.
Chao Gong, Dong Li, Yingwei Pan, Jingjing Chen, Ting Yao, Tao Mei.
AAAI 2026. [PDF] [Code]

Reversible Inversion for Training-Free Exemplar-guided Image Editing.
Yuke Li, Lianli Gao, Ji Zhang, Pengpeng Zeng, Lichuan Xiang, Hongkai Wen, Heng Tao Shen, Jingkuan Song.
arxiv 2025. [PDF] [Code]

FreeControl: Efficient, Training-Free Structural Control via One-Step Attention Extraction.
Jiang Lin, Xinyu Chen, Song Wu, Zhiqiu Zhang, Jizhi Zhang, Ye Wang, Qiang Tang, Qian Wang, Jian Yang, Zili Yi.
NeurIPS 2025. [PDF]

SplitFlow: Flow Decomposition for Inversion-Free Text-to-Image Editing.
Sung-Hoon Yoon, Minghan Li, Gaspard Beaudouin, Congcong Wen, Muhammad Rafay Azhar, Mengyu Wang.
NeurIPS 2025. [PDF] [Code]

RegionE: Adaptive Region-Aware Generation for Efficient Image Editing.
Pengtao Chen, Xianfang Zeng, Maosen Zhao, Mingzhu Shen, Peng Ye, Bangyin Xiang, Zhibo Wang, Wei Cheng, Gang Yu, Tao Chen.
arxiv 2025. [PDF] [Code]

OmniText: A Training-Free Generalist for Controllable Text-Image Manipulation.
Agus Gunawan, Samuel Teodoro, Yun Chen, Soo Ye Kim, Jihyong Oh, Munchurl Kim.
arxiv 2025. [PDF] [Project]

ConsistEdit: Highly Consistent and Precise Training-free Visual Editing.
Zixin Yin, Ling-Hao Chen, Lionel Ni, Xili Dai.
SIGGRAPH Asia 2025. [PDF] [Project] [Code]

Coupled Diffusion Sampling for Training-Free Multi-View Image Editing.
Hadi Alzayer, Yunzhi Zhang, Chen Geng, Jia-Bin Huang, Jiajun Wu.
arxiv 2025. [PDF] [Project] [Code]

RetouchLLM: Training-free White-box Image Retouching.
Moon Ye-Bin, Roy Miles, Tae-Hyun Oh, Ismail Elezi, Jiankang Deng.
arxiv 2025. [PDF]

Teleportraits: Training-Free People Insertion into Any Scene.
Jialu Gao, K J Joseph, Fernando De La Torre.
arxiv 2025. [PDF]

Token Painter: Training-Free Text-Guided Image Inpainting via Mask Autoregressive Models.
Longtao Jiang, Mingfei Han, Lei Chen, Yongqiang Yu, Feng Zhao, Xiaojun Chang, Zhihui Li.
arxiv 2025. [PDF]

FreeInsert: Personalized Object Insertion with Geometric and Style Control.
Yuhong Zhang, Han Wang, Yiwen Wang, Rong Xie, Li Song.
arxiv 2025. [PDF]

Training-Free Text-Guided Color Editing with Multi-Modal Diffusion Transformer.
Zixin Yin, Xili Dai, Ling-Hao Chen, Deyu Zhou, Jianan Wang, Duomin Wang, Gang Yu, Lionel M. Ni, Heung-Yeung Shum.
arxiv 2025. [PDF] [Project]

CannyEdit: Selective Canny Control and Dual-Prompt Guidance for Training-Free Image Editing.
Weiyan Xie, Han Gao, Didan Deng, Kaican Li, April Hua Liu, Yongxiang Huang, Nevin L. Zhang.
arxiv 2025. [PDF] [Project]

UniEdit-I: Training-free Image Editing for Unified VLM via Iterative Understanding, Editing and Verifying.
Chengyu Bai, Jintao Chen, Xiang Bai, Yilong Chen, Qi She, Ming Lu, Shanghang Zhang.
arxiv 2025. [PDF]

Training-free Geometric Image Editing on Diffusion Models.
Hanshen Zhu, Zhen Zhu, Kaile Zhang, Yiming Gong, Yuliang Liu, Xiang Bai.
ICCV 2025. [PDF] [Code]

OmniVTON: Training-Free Universal Virtual Try-On.
Zhaotong Yang, Yuhui Li, Shengfeng He, Xinzhe Li, Yangyang Xu, Junyu Dong, Yong Du.
ICCV 2025. [PDF] [Code]

Towards Generalized and Training-Free Text-Guided Semantic Manipulation.
Yu Hong, Xiao Cai, Pengpeng Zeng, Shuai Zhang, Jingkuan Song, Lianli Gao, Heng Tao Shen.
arxiv 2025. [PDF]

Anchor Token Matching: Implicit Structure Locking for Training-free AR Image Editing.
Taihang Hu, Linxuan Li, Kai Wang, Yaxing Wang, Jian Yang, Ming-Ming Cheng.
arxiv 2025. [PDF] [Code]

Tuning-Free Image Editing with Fidelity and Editability via Unified Latent Diffusion Model.
Qi Mao, Lan Chen, Yuchao Gu, Mike Zheng Shou, Ming-Hsuan Yang.
arxiv 2025. [PDF] [Code]

Training-Free Text-Guided Image Editing with Visual Autoregressive Model.
Yufei Wang, Lanqing Guo, Zhihao Li, Jiaxing Huang, Pichao Wang, Bihan Wen, Jian Wang.
arxiv 2025. [PDF] [Code]

OmnimatteZero: Training-free Real-time Omnimatte with Pre-trained Video Diffusion Models.
Dvir Samuel, Matan Levy, Nir Darshan, Gal Chechik, Rami Ben-Ari.
arxiv 2025. [PDF]

FreeFlux: Understanding and Exploiting Layer-Specific Roles in RoPE-Based MMDiT for Versatile Image Editing.
Tianyi Wei, Yifan Zhou, Dongdong Chen, Xingang Pan.
arxiv 2025. [PDF] [Project] [Code]

NullFace: Training-Free Localized Face Anonymization.
Han-Wei Kung, Tuomas Varanka, Terence Sim, Nicu Sebe.
arxiv 2025. [PDF] [Code]

PTDiffusion: Free Lunch for Generating Optical Illusion Hidden Pictures with Phase-Transferred Diffusion Model.
Xiang Gao, Shuai Yang, Jiaying Liu.
CVPR 2025. [PDF]

KV-Edit: Training-Free Image Editing for Precise Background Preservation.
Tianrui Zhu, Shiyi Zhang, Jiawei Shao, Yansong Tang.
arxiv 2025. [PDF] [Project] [Code]

360PanT: Training-Free Text-Driven 360-Degree Panorama-to-Panorama Translation.
Hai Wang, Jing-Hao Xue.
WACV 2025. [PDF] [Project] [Code]

OmniSSR: Zero-shot Omnidirectional Image Super-Resolution using Stable Diffusion Model.
Runyi Li, Xuhan Sheng, Weiqi Li, Jian Zhang.
ECCV 2024. [PDF] [Project] [Code]

Guide-and-Rescale: Self-Guidance Mechanism for Effective Tuning-Free Real Image Editing.
Vadim Titov, Madina Khalmatova, Alexandra Ivanova, Dmitry Vetrov, Aibek Alanov.
arxiv 2024. [PDF] [Code]

TALE: Training-free Cross-domain Image Composition via Adaptive Latent Manipulation and Energy-guided Optimization.
Kien T. Pham, Jingye Chen, Qifeng Chen.
ACM MM 2024. [PDF] [Project] [Code]

Faster Diffusion via Temporal Attention Decomposition.
Haozhe Liu, Wentian Zhang, Jinheng Xie, Francesco Faccio, Mengmeng Xu, Tao Xiang, Mike Zheng Shou, Juan-Manuel Perez-Rua, JΓΌrgen Schmidhuber.
arXiv 2024. [PDF] [Code]

DiffUHaul: A Training-Free Method for Object Dragging in Images.
Omri Avrahami, Rinon Gal, Gal Chechik, Ohad Fried, Dani Lischinski, Arash Vahdat, Weili Nie.
Siggraph Asia 2024. [PDF] [Project]

Analogist: Out-of-the-box Visual In-Context Learning with Image Diffusion Model.
Zheng Gu, Shiyuan Yang, Jing Liao, Jing Huo, Yang Gao.
Siggraph 2024. [PDF] [Project] [Code]

ObjectAdd: Adding Objects into Image via a Training-Free Diffusion Modification Fashion.
Ziyue Zhang, Mingbao Lin, Rongrong Ji.
arXiv 2024. [PDF]

CutDiffusion: A Simple, Fast, Cheap, and Strong Diffusion Extrapolation Method.
Mingbao Lin, Zhihang Lin, Wengyi Zhan, Liujuan Cao, Rongrong Ji.
arXiv 2024. [PDF] [Project]

FreeDiff: Progressive Frequency Truncation for Image Editing with Diffusion Models.
Wei Wu, Qingnan Fan, Shuai Qin, Hong Gu, Ruoyu Zhao, Antoni B. Chan.
arXiv 2024. [PDF]

FreeControl: Training-Free Spatial Control of Any Text-to-Image Diffusion Model with Any Condition.
Sicheng Mo, Fangzhou Mu, Kuan Heng Lin, Yanli Liu, Bochen Guan, Yin Li, Bolei Zhou.
CVPR 2024. [PDF] [Project] [Code]

Plug-and-Play Diffusion Features for Text-Driven Image-to-Image Translation.
Narek Tumanyan, Michal Geyer, Shai Bagon, Tali Dekel.
CVPR 2023. [PDF] [Project] [Code]

πŸ” Return to Top

Style Transfer

SceneTextStylizer: A Training-Free Scene Text Style Transfer Framework with Diffusion Model.
Honghui Yuan, Keiji Yanai.
arxiv 2025. [PDF]

Training-Free Multi-Style Fusion Through Reference-Based Adaptive Modulation.
Xu Liu, Yibo Lu, Xinxian Wang, Xinyu Wu.
ACPR 2025. [PDF]

Training-Free Identity Preservation in Stylized Image Generation Using Diffusion Models.
Mohammad Ali Rezaei, Helia Hajikazem, Saeed Khanehgir, Mahdi Javanmardi.
arxiv 2025. [PDF]

Training Free Stylized Abstraction.
Aimon Rahman, Kartik Narayan, Vishal M. Patel.
arxiv 2025. [PDF] [Project] [Code]

Training-free Stylized Text-to-Image Generation with Fast Inference.
Xin Ma, Yaohui Wang, Xinyuan Chen, Tien-Tsin Wong, Cunjian Chen.
arxiv 2025. [PDF] [Project] [Code]

DriveGEN: Generalized and Robust 3D Detection in Driving via Controllable Text-to-Image Diffusion Generation.
Hongbin Lin, Zilu Guo, Yifan Zhang, Shuaicheng Niu, Yafeng Li, Ruimao Zhang, Shuguang Cui, Zhen Li.
CVPR 2025. [PDF] [Code]

Free-Lunch Color-Texture Disentanglement for Stylized Image Generation.
Jiang Qin, Senmao Li, Alexandra Gomez-Villa, Shiqi Yang, Yaxing Wang, Kai Wang, Joost van de Weijer.
arXiv 2025. [PDF]

AttenST: A Training-Free Attention-Driven Style Transfer Framework with Pre-Trained Diffusion Models.
Bo Huang, Wenlun Xu, Qizhuo Han, Haodong Jing, Ying Li.
arXiv 2025. [PDF]

K-LoRA: Unlocking Training-Free Fusion of Any Subject and Style LoRAs.
Ziheng Ouyang, Zhen Li, Qibin Hou.
arXiv 2025. [PDF]

Artist: Aesthetically Controllable Text-Driven Stylization without Training.
Ruixiang Jiang, Changwen Chen.
arXiv 2024. [PDF] [Project] [Code]

Visual Style Prompting with Swapping Self-Attention.
Jaeseok Jeong, Junho Kim, Yunjey Choi, Gayoung Lee, Youngjung Uh.
arXiv 2024. [PDF] [Project] [Code]

FreeStyle: Free Lunch for Text-guided Style Transfer using Diffusion Models.
Feihong He, Gang Li, Mengyuan Zhang, Leilei Yan, Lingyu Si, Fanzhang Li.
arXiv 2024. [PDF] [Project] [Code]

Eye-for-an-eye: Appearance Transfer with Semantic Correspondence in Diffusion Models.
Sooyeon Go, Kyungmook Choi, Minjung Shin, Youngjung Uh.
arXiv 2024. [PDF] [Project] [Code]

Ctrl-X: Controlling Structure and Appearance for Text-To-Image Generation Without Guidance.
Kuan Heng Lin, Sicheng Mo, Ben Klingher, Fangzhou Mu, Bolei Zhou.
NeurIPS 2024. [PDF] [Project] [Code]

RB-Modulation: Training-Free Personalization of Diffusion Models using Stochastic Optimal Control.
Litu Rout, Yujia Chen, Nataniel Ruiz, Abhishek Kumar, Constantine Caramanis, Sanjay Shakkottai, Wen-Sheng Chu.
arXiv 2024. [PDF] [Project] [Code]

Tuning-Free Adaptive Style Incorporation for Structure-Consistent Text-Driven Style Transfer.
Yanqi Ge, Jiaqi Liu, Qingnan Fan, Xi Jiang, Ye Huang, Shuai Qin, Hong Gu, Wen Li, Lixin Duan.
arXiv 2024. [PDF]

Cross-Image Attention for Zero-Shot Appearance Transfer.
Yuval Alaluf, Daniel Garibi, Or Patashnik, Hadar Averbuch-Elor, Daniel Cohen-Or.
SIGGRAPH 2024. [PDF] [Project] [Code]

Style Injection in Diffusion: A Training-free Approach for Adapting Large-scale Diffusion Models for Style Transfer.
Jiwoo Chung, Sangeek Hyun, Jae-Pil Heo.
CVPR 2024. [PDF] [Project] [Code]

πŸ” Return to Top

Material Transfer

ZeST: Zero-Shot Material Transfer from a Single Image.
Ta-Ying Cheng, Prafull Sharma, Andrew Markham, Niki Trigoni, Varun Jampani.
ECCV 2024. [PDF] [Project] [Code]

πŸ” Return to Top

Video Generation

PackCache: A Training-Free Acceleration Method for Unified Autoregressive Video Generation via Compact KV-Cache.
Kunyang Li, Mubarak Shah, Yuzhang Shang.
arxiv 2026. [PDF]

CineLOG: A Training Free Approach for Cinematic Long Video Generation.
Zahra Dehghanian, Morteza Abolghasemi, Hamid Beigy, Hamid R. Rabiee.
arxiv 2025. [PDF] [Project]

Deep Forcing: Training-Free Long Video Generation with Deep Sink and Participative Compression.
Jung Yi, Wooseok Jang, Paul Hyunbin Cho, Jisu Nam, Heeji Yoon, Seungryong Kim.
arxiv 2025. [PDF] [Project] [Code]

AlignVid: Training-Free Attention Scaling for Semantic Fidelity in Text-Guided Image-to-Video Generation.
Yexin Liu, Wen-Jie Shu, Zile Huang, Haoze Zheng, Yueze Wang, Manyuan Zhang, Ser-Nam Lim, Harry Yang.
arxiv 2025. [PDF]

Block Cascading: Training Free Acceleration of Block-Causal Video Models.
Hmrishav Bandyopadhyay, Nikhil Pinnaparaju, Rahim Entezari, Jim Scott, Yi-Zhe Song, Varun Jampani.
arxiv 2025. [PDF] [Project]

FreeSwim: Revisiting Sliding-Window Attention Mechanisms for Training-Free Ultra-High-Resolution Video Generation.
Yunfeng Wu, Jiayi Song, Zhenxiong Tan, Zihao He, Songhua Liu.
arxiv 2025. [PDF] [Code]

BachVid: Training-Free Video Generation with Consistent Background and Character.
Han Yan, Xibin Song, Yifu Wang, Hongdong Li, Pan Ji, Chao Ma.
arxiv 2025. [PDF] [Project] [Code]

Playmate2: Training-Free Multi-Character Audio-Driven Animation via Diffusion Transformer with Reward Feedback.
Xingpei Ma, Shenneng Huang, Jiaran Cai, Yuansheng Guan, Shen Zheng, Hanfeng Zhao, Qiang Zhang, Shunsi Zhang.
arxiv 2025. [PDF] [Project] [Code]

LightCache: Memory-Efficient, Training-Free Acceleration for Video Generation.
Yang Xiao, Gen Li, Kaiyuan Deng, Yushu Wu, Zheng Zhan, Yanzhi Wang, Xiaolong Ma, Bo Hui.
arxiv 2025. [PDF] [Code]

DiTraj: training-free trajectory control for video diffusion transformer.
Cheng Lei, Jiayu Zhang, Yue Ma, Xinyu Wang, Long Chen, Liang Tang, Yiqiang Yan, Fei Su, Zhicheng Zhao.
arxiv 2025. [PDF]

Identity-Preserving Text-to-Video Generation via Training-Free Prompt, Image, and Guidance Enhancement.
Jiayi Gao, Changcheng Hua, Qingchao Chen, Yuxin Peng, Yang Liu.
arxiv 2025. [PDF] [Code]

Less is Enough: Training-Free Video Diffusion Acceleration via Runtime-Adaptive Caching.
Xin Zhou, Dingkang Liang, Kaijin Chen, Tianrui Feng, Xiwu Chen, Hongkai Lin, Yikang Ding, Feiyang Tan, Hengshuang Zhao, Xiang Bai.
arxiv 2025. [PDF] [Project] [Code]

FreeLong++: Training-Free Long Video Generation via Multi-band SpectralFusion.
Yu Lu, Yi Yang.
arxiv 2025. [PDF] [Project] [Code]

Training-Free Motion Customization for Distilled Video Generators with Adaptive Test-Time Distillation.
Jintao Rong, Xin Xie, Xinyi Yu, Linlin Ou, Xinyu Zhang, Chunhua Shen, Dong Gong.
arxiv 2025. [PDF] [Project]

Frame Guidance: Training-Free Guidance for Frame-Level Control in Video Diffusion Models.
Sangwon Jang, Taekyung Ki, Jaehyeong Jo, Jaehong Yoon, Soo Ye Kim, Zhe Lin, Sung Ju Hwang.
arxiv 2025. [PDF] [Project] [Code]

DiffuseSlide: Training-Free High Frame Rate Video Generation Diffusion.
Geunmin Hwang, Hyun-kyu Ko, Younghyun Kim, Seungryong Lee, Eunbyung Park.
arxiv 2025. [PDF] [Project] [Code]

MOVi: Training-free Text-conditioned Multi-Object Video Generation.
Aimon Rahman, Jiang Liu, Ze Wang, Ximeng Sun, Jialian Wu, Xiaodong Yu, Yusheng Su, Vishal M. Patel, Zicheng Liu, Emad Barsoum.
arxiv 2025. [PDF] [Code]

Training-Free Efficient Video Generation via Dynamic Token Carving.
Yuechen Zhang, Jinbo Xing, Bin Xia, Shaoteng Liu, Bohao Peng, Xin Tao, Pengfei Wan, Eric Lo, Jiaya Jia.
arxiv 2025. [PDF] [Project] [Code]

DynamicScaler: Seamless and Scalable Video Generation for Panoramic Scenes.
Jinxiu Liu, Shaoheng Lin, Yinxiao Li, Ming-Hsuan Yang.
CVPR 2025. [PDF] [Project] [Code]

FreePCA: Integrating Consistency Information across Long-short Frames in Training-free Long Video Generation via Principal Component Analysis.
Jiangtong Tan, Hu Yu, Jie Huang, Jie Xiao, Feng Zhao.
CVPR 2025. [PDF] [Code]

Training-free Guidance in Text-to-Video Generation via Multimodal Planning and Structured Noise Initialization.
Jialu Li, Shoubin Yu, Han Lin, Jaemin Cho, Jaehong Yoon, Mohit Bansal.
arxiv 2025. [PDF] [Project] [Code]

EIDT-V: Exploiting Intersections in Diffusion Trajectories for Model-Agnostic, Zero-Shot, Training-Free Text-to-Video Generation.
Diljeet Jagpal, Xi Chen, Vinay P. Namboodiri.
CVPR 2025. [PDF] [Project] [Code]

Every Painting Awakened: A Training-free Framework for Painting-to-Animation Generation.
Lingyu Liu, Yaxiong Wang, Li Zhu, Zhedong Zheng.
arxiv 2025. [PDF] [Project]

On-device Sora: Enabling Training-Free Diffusion-based Text-to-Video Generation for Mobile Devices.
Bosung Kim, Kyuhwan Lee, Isu Jeong, Jungmin Cheon, Yeojin Lee, Seulki Lee.
arxiv 2025. [PDF] [Code]

Zero4D: Training-Free 4D Video Generation From Single Video Using Off-the-Shelf Video Diffusion Model.
Jangho Park, Taesung Kwon, Jong Chul Ye.
arxiv 2025. [PDF] [Project]

MagicComp: Training-free Dual-Phase Refinement for Compositional Video Generation.
Hongyu Zhang, Yufan Deng, Shenghai Yuan, Peng Jin, Zesen Cheng, Yian Zhao, Chang Liu, Jie Chen.
arxiv 2025. [PDF] [Project] [Code]

MotionMaster: Training-free Camera Motion Transfer For Video Generation.
Teng Hu, Jiangning Zhang, Ran Yi, Yating Wang, Hongrui Huang, Jieyu Weng, Yabiao Wang, Lizhuang Ma.
ACM MM 2024. [PDF] [Code]

πŸ” Return to Top

Video Manipulation

Object-WIPER : Training-Free Object and Associated Effect Removal in Videos.
Saksham Singh Kushwaha, Sayan Nag, Yapeng Tian, Kuldeep Kulkarni.
arxiv 2026. [PDF] [Project]

FAME: Fairness-aware Attention-modulated Video Editing.
Zhangkai Wu, Xuhui Fan, Zhongyuan Xie, Kaize Shi, Zhidong Li, Longbing Cao.
arxiv 2025. [PDF]

VALA: Learning Latent Anchors for Training-Free and Temporally Consistent Video Editing.
Zhangkai Wu, Xuhui Fan, Zhongyuan Xie, Kaize Shi, Longbing Cao.
arxiv 2025. [PDF]

ConsistEdit: Highly Consistent and Precise Training-free Visual Editing.
Zixin Yin, Ling-Hao Chen, Lionel Ni, Xili Dai.
SIGGRAPH Asia 2025. [PDF] [Project] [Code]

FreeViS: Training-free Video Stylization with Inconsistent References.
Jiacong Xu, Yiqun Mei, Ke Zhang, Vishal M. Patel.
arxiv 2025. [PDF] [Project] [Code]

ContextFlow: Training-Free Video Object Editing via Adaptive Context Enrichment.
Yiyang Chen, Xuanhua He, Xiujun Ma, Yue Ma.
arxiv 2025. [PDF] [Project] [Code]

STR-Match: Matching SpatioTemporal Relevance Score for Training-Free Video Editing.
Junsung Lee, Junoh Kang, Bohyung Han.
arxiv 2025. [PDF] [Project] [Code]

Good Noise Makes Good Edits: A Training-Free Diffusion-Based Video Editing with Image and Text Prompts.
Saemee Choi, Sohyun Jeong, Jaegul Choo, Jinhee Kim.
arxiv 2025. [PDF]

TV-LiVE: Training-Free, Text-Guided Video Editing via Layer Informed Vitality Exploitation.
Min-Jung Kim, Dongjin Kim, Seokju Yun, Jaegul Choo.
arxiv 2025. [PDF] [Project]

FlowDirector: Training-Free Flow Steering for Precise Text-to-Video Editing.
Guangzhao Li, Yanming Yang, Chenxi Song, Chi Zhang.
arxiv 2025. [PDF] [Project] [Code]

RoPECraft: Training-Free Motion Transfer with Trajectory-Guided RoPE Optimization on Diffusion Transformers.
Ahmet Berke Gokmen, Yigit Ekin, Bahri Batuhan Bilecen, Aysegul Dundar.
arxiv 2025. [PDF] [Project] [Code]

RASA: Replace Anyone, Say Anything -- A Training-Free Framework for Audio-Driven and Universal Portrait Video Editing.
Tianrui Pan, Lin Liu, Jie Liu, Xiaopeng Zhang, Jie Tang, Gangshan Wu, Qi Tian.
arxiv 2025. [PDF] [Project]

Light-A-Video: Training-free Video Relighting via Progressive Light Fusion.
Yujie Zhou, Jiazi Bu, Pengyang Ling, Pan Zhang, Tong Wu, Qidong Huang, Jinsong Li, Xiaoyi Dong, Yuhang Zang, Yuhang Cao, Anyi Rao, Jiaqi Wang, Li Niu.
arxiv 2025. [PDF] [Project] [Code]

Slicedit: Zero-Shot Video Editing With Text-to-Image Diffusion Models Using Spatio-Temporal Slices.
Nathaniel Cohen, Vladimir Kulikov, Matan Kleiner, Inbar Huberman-Spiegelglas, Tomer Michaeli.
ICML 2024. [PDF] [Project] [Code]

FLATTEN: optical FLow-guided ATTENtion for consistent text-to-video editing.
Yuren Cong, Mengmeng Xu, Christian Simon, Shoufa Chen, Jiawei Ren, Yanping Xie, Juan-Manuel Perez-Rua, Bodo Rosenhahn, Tao Xiang, Sen He.
ICLR 2024. [PDF] [Project] [Code]

TokenFlow: Consistent Diffusion Features for Consistent Video Editing.
Michal Geyer, Omer Bar-Tal, Shai Bagon, Tali Dekel.
ICLR 2024. [PDF] [Project] [Code]

πŸ” Return to Top

3D Generation

Fast3Dcache: Training-free 3D Geometry Synthesis Acceleration.
Mengyu Yang, Yanming Yang, Chenyi Xu, Chenxi Song, Yufan Zuo, Tong Zhao, Ruibo Li, Chi Zhang.
arxiv 2025. [PDF] [Project] [Code]

FreeArt3D: Training-Free Articulated Object Generation using 3D Diffusion.
Chuhao Chen, Isabella Liu, Xinyue Wei, Hao Su, Minghua Liu.
SIGGRAPH Asia 2025. [PDF] [Project] [Code]

TRELLISWorld: Training-Free World Generation from Object Generators.
Hanke Chen, Yuan Liu, Minchen Li.
arxiv 2025. [PDF]

πŸ” Return to Top

3D Manipulation

AnchorFlow: Training-Free 3D Editing via Latent Anchor-Aligned Flows.
Zhenglin Zhou, Fan Ma, Chengzhuo Gui, Xiaobo Xia, Hehe Fan, Yi Yang, Tat-Seng Chua.
arxiv 2025. [PDF] [Code]

NANO3D: A Training-Free Approach for Efficient 3D Editing Without Masks.
Junliang Ye, Shenghao Xie, Ruowen Zhao, Zhengyi Wang, Hongyu Yan, Wenqiang Zu, Lei Ma, Jun Zhu.
arxiv 2025. [PDF] [Project] [Code]

πŸ” Return to Top

4D Manipulation

Dynamic-eDiTor: Training-Free Text-Driven 4D Scene Editing with Multimodal Diffusion Transformer.
Dong In Lee, Hyungjun Doh, Seunggeun Chi, Runlin Duan, Sangpil Kim, Karthik Ramani.
arxiv 2025. [PDF] [Project]

πŸ” Return to Top

About

This is a repository to collect training-free algorithms for visual generation and manipulation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published