UniTune: Text-Driven Image Editing by Fine Tuning an Image Generation Model on a Single Image
We present unitune, a simple and novel method for general text-driven imageediting.
Unitune gets as input an arbitrary image and a textual editdescription, and carries out the edit while maintaining high semantic and visual fidelity to the input image.
At the core of our method is the observation that with the rightchoice of parameters, we can fine-tune a large text-to-image diffusion model on a single image, encouraging the model to maintain fidelity to the input image while still allowing expressive manipulations.
We test our method in a range of different use cases, and demonstrate its wide applicability.