- Notifications
You must be signed in to change notification settings - Fork57
MinImagen: A minimal implementation of the Imagen text-to-image model
License
AssemblyAI-Community/MinImagen
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
A Minimal implementation of theImagen text-to-image model.
SeeBuild Your Own Imagen Text-to-Image Model for a tutorial on how to build MinImagen.
SeeHow Imagen Actually Works for a detailed explanation of Imagen's operating principles.
Given a caption of an image, the text-to-image modelImagen will generate an image that reflects the scene described by the caption. The model is acascading diffusion model, using aT5 text encoder to generate a caption encoding which conditions a base image generator and then a sequence of super-resolution models through which the output of the base image generator is passed.
In particular, two notable contributions are the developments of:
- Noise Conditioning Augmentation, which noises low-resolution conditioning images in the super-resolution models, and
- Dynamic Thresholding which helps prevent image saturation at highclassifier-free guidance weights.
N.B. - This project is intended only for educational purposes to demonstrate how Diffusion Models are implemented and incorporated into text-to-image models. Many components of the network that are not essential for these educational purposes have been stripped off for simplicity. For a full-fledged implementation, check out Phil Wang's repo (see attribution note below)
- Attribution Note
- Installation
- Documentation
- Usage - Command Line
main.py
- training and image generation in sequencetrain.py
- training a MinImagen instanceinference.py
- generated images using a MinImagen instance
- Usage - Package
- Modifying the Source Code
- Additional Resources
- Socials
This implementation is largely based on Phil Wang'sImagen implementation.
To install MinImagen, run the following command in the terminal:
$ pip install minimagen
Note that MinImagen requires Python3.9 or higher
See theMinImagen Documentation to learn more about the package.
If you have cloned this repo (as opposed to just installing theminimagen
package), you can use the provided scripts to get started with MinImagen. This repo can be cloned by running the following command in the terminal:
$ git clone https://github.com/AssemblyAI-Examples/MinImagen.git
For the most basic usage, simply enter the MinImagen directory and run the following in the terminal:
$ python main.py
This will create a small MinImagen instance and train it on a tiny amount of data, and then use this MinImagen instance to generate an image.
After running the script, you will see a directory calledtraining_<TIMESTAMP>
.
- This directory is called aTraining Directory and is generated when training a MinImagen instance.
- It contains information about the configuration (
parameters
subdirectory), and contains the model checkpoints (state_dicts
andtmp
directories). - It also contains a
training_progress.txt
file that records training progress.
You will also see a directory calledgenerated_images_<TIMESTEP>
.
- This directory contains a folder of images generated by the model (
generated_images
). - It also contains
captions.txt
files, which documents the captions that were input to get the images (where the line index of a given caption corresponds to the image number in thegenerated_iamges
folder). - Finally, this directory also contains
imagen_training_directory.txt
, which specifies the name of the Training Directory used to load the MinImagen instance / generate images.
main.py
simply runstrain.py
andinference.py
in series, the former to train the model and the latter to generate the image.
To train a model, simply runtrain.py
and specify relevant command line arguments. Thepossible arguments are:
--PARAMETERS
or-p
, which specifies a directory that specifies the MinImagen configuration to use. It should be structured like aparameters
subdirectory within a Training Directory (example inparameters
).--NUM_WORKERS"
or-n
, which specifies the number of workers to use for the DataLoaders.--BATCH_SIZE
or-b
, which specifies the batch size to use during training.--MAX_NUM_WORDS
or-mw
, which specifies the maximum number of words allowed in a caption.--IMG_SIDE_LEN
or-s
, specifies the final side length of the square images the MinImagen will output.--EPOCHS
or-e
, which specifies the number of training epochs.--T5_NAME
-t5
, which specifies the name of T5 encoder to use.--TRAIN_VALID_FRAC
or-f
, which specifies the fraction of dataset to use for training (vs. validation).--TIMESTEPS
or-t
, which specifies the number of timesteps in Diffusion Process.--OPTIM_LR
or-lr
, which specifies the learning rate for Adam optimizer.--ACCUM_ITER
or-ai
, which specifies the number of batches to accumulate for gradient accumulation.--CHCKPT_NUM
or-cn
, which specifies the interval of batches to create a temporary model checkpoint at during training.--VALID_NUM
or-vn
, which specifies the number of validation images to use. If None, uses full amount from train/valid split. The reason for including this is that, even with an e.g. 0.99--TRAIN_VALID_FRAC
, a prohibitively large number of images could still be left for validation for very large datasets.--RESTART_DIRECTORY
or-rd
, training directory to load MinImagen instance from if resuming training. A new Training Directory will be created for the training, leaving the previous Training Directory from which the checkpoint is loaded unperturbed.--TESTING
or-test
, which is used to run the script with a small MinImagen instance and small dataset for testing.
For example, to run a small training using the provided exampleparameters
folder, run the following in the terminal:
python train.py --PARAMETERS ./parameters --BATCH_SIZE 2 --TIMESTEPS 25 --TESTING
After execution, you will see a newtraining_<TIMESTAMP>
Training Directory that contains the files aslisted above from the training.
To generate images using a model from aTraining Directory, we can useinference.py
. Simply runinference.py
and specify relevant command line arguments. The possible arguments are:
--TRAINING_DIRECTORY"
or-d
, which specifies the training directory from which to load the MinImagen instance for inference.--CAPTIONS
or-c
, which specifies either (a) a single caption to generate an image for, or (b) a filepath to a.txt
file that contains a list of captions to generate images for, where each caption is on a new line.
For example, to generate images for the example captions provided incaptions.txt
using the model generated from the above training line, simply run
python inference.py -CAPTIONS captions.txt --TRAINING_DIRECTORY training_<TIMESTAMP>
whereTIMESTAMP
is replaced with the appropriate value from your training.
A minimal training script using theminimagen
package is shown below. Seetrain.py
for a more built-up version of the below code.
importosfromdatetimeimportdatetimeimporttorch.utils.datafromtorchimportoptimfromminimagen.ImagenimportImagenfromminimagen.UnetimportUnet,Base,Super,BaseTest,SuperTestfromminimagen.generateimportload_minimagen,load_paramsfromminimagen.t5importget_encoded_dimfromminimagen.trainingimportget_minimagen_parser,ConceptualCaptions,get_minimagen_dl_opts, \create_directory,get_model_size,save_training_info,get_default_args,MinimagenTrain, \load_testing_parameters# Get devicedevice=torch.device("cuda:0"iftorch.cuda.is_available()else"cpu")# Command line argument parserparser=get_minimagen_parser()args=parser.parse_args()# Create training directorytimestamp=datetime.now().strftime("%Y%m%d_%H%M%S")dir_path=f"./training_{timestamp}"training_dir=create_directory(dir_path)# Replace some cmd line args to lower computational load.args=load_testing_parameters(args)# Load subset of Conceptual Captions dataset.train_dataset,valid_dataset=ConceptualCaptions(args,smalldata=True)# Create dataloadersdl_opts= {**get_minimagen_dl_opts(device),'batch_size':args.BATCH_SIZE,'num_workers':args.NUM_WORKERS}train_dataloader=torch.utils.data.DataLoader(train_dataset,**dl_opts)valid_dataloader=torch.utils.data.DataLoader(valid_dataset,**dl_opts)# Use small U-Nets to lower computational load.unets_params= [get_default_args(BaseTest),get_default_args(SuperTest)]unets= [Unet(**unet_params).to(device)forunet_paramsinunets_params]# Specify MinImagen parametersimagen_params=dict(image_sizes=(int(args.IMG_SIDE_LEN/2),args.IMG_SIDE_LEN),timesteps=args.TIMESTEPS,cond_drop_prob=0.15,text_encoder_name=args.T5_NAME)# Create MinImagen from UNets with specified imagen parametersimagen=Imagen(unets=unets,**imagen_params).to(device)# Fill in unspecified arguments with defaults to record complete config (parameters) fileunets_params= [{**get_default_args(Unet),**i}foriinunets_params]imagen_params= {**get_default_args(Imagen),**imagen_params}# Get the size of the Imagen model in megabytesmodel_size_MB=get_model_size(imagen)# Save all training info (config files, model size, etc.)save_training_info(args,timestamp,unets_params,imagen_params,model_size_MB,training_dir)# Create optimizeroptimizer=optim.Adam(imagen.parameters(),lr=args.OPTIM_LR)# Train the MinImagen instanceMinimagenTrain(timestamp,args,unets,imagen,train_dataloader,valid_dataloader,training_dir,optimizer,timeout=30)
A minimal inference script using theminimagen
package is shown below. Seeinference.py
for a more built-up version of the below code.
fromargparseimportArgumentParserfromminimagen.generateimportload_minimagen,sample_and_save# Command line argument parserparser=ArgumentParser()parser.add_argument("-d","--TRAINING_DIRECTORY",dest="TRAINING_DIRECTORY",help="Training directory to use for inference",type=str)args=parser.parse_args()# Specify the caption(s) to generate images forcaptions= ['a happy dog']# Use `sample_and_save` to generate and save the iamgessample_and_save(captions,training_directory=args.TRAINING_DIRECTORY)# Alternatively, rather than specifying a Training Directory, you can input just a MinImagen instance to use for image generation.# In this case, information about the MinImagen instance used to generate the images will not be saved.minimagen=load_minimagen(args.TRAINING_DIRECTORY)sample_and_save(captions,minimagen=minimagen)
To see more of what MinImagen has to offer, or to get additional details on the scripts above, check out theMinImagen Documentation
If you want to make modifications to the source code (rather than use theminimagen
package), first clone this repository and navigate into it:
$ git clone https://github.com/AssemblyAI-Examples/MinImagen.git$cd MinImagen
After that, create a virtual environment:
$ pip install virtualenv$ virtualenv venv
Then activate the virtual environment and install all dependencies:
$ .\venv\Scripts\activate.bat# Windows$source venv/bin/activate# MacOS/Linux$ pip install -r requirements.txt
Now you can modify the source code and the changes will be reflected when running any of theincluded scripts (as long as the virtual environment created above is active).
- For a step-by-step guide on how to build the version of Imagen in this repository, seeBuild Your Own Imagen Text-to-Image Model.
- For an deep-dive into how Imagen works, seeHow Imagen Actually Works.
- For a deep-dive into Diffusion Models, see ourIntroduction to Diffusion Models for Machine Learning guide.
- For additional learning resources on Machine Learning and Deep Learning, check out ourBlog andYouTube channel.
- Read the original Imagen paperhere.
- Follow us onTwitter for more Deep Learning content.
- Follow our newsletter to stay up to date on our recent content.