Skip to content

This is a PyTorch-based implementation of the Generative Adversarial Text-to-Image Synthesis paper, utilizing a GAN architecture inspired by DCGAN with text descriptions as inputs to generate images.

Notifications You must be signed in to change notification settings

yashashwita20/text-to-image-using-GAN

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Text to Image using DCGAN

This implementation is a PyTorch-based version of Generative Adversarial Text-to-Image Synthesis paper. In this project, a Conditional Generative Adversarial Network (CGAN) is trained, leveraging text descriptions as conditioning inputs to generate corresponding images. The architecture of this model draws inspiration from DCGAN (Deep Convolutional Generative Adversarial Network).

Requirements

  • h5py==3.6.0
  • numpy==1.21.5
  • Pillow==10.0.0
  • torch==2.0.0

Dataset

We used Caltech-UCSD Birds 200 and text embeddings provided by Reed Scott et al.

Repository

├── models
├     └──  dcgan_model.py
├── utils.py
├── data_util.py
├── requirements.txt
└──  DCGAN_Text2Image.ipynb

Results

References

[1] Generative Adversarial Text-to-Image Synthesis https://arxiv.org/abs/1605.05396

About

This is a PyTorch-based implementation of the Generative Adversarial Text-to-Image Synthesis paper, utilizing a GAN architecture inspired by DCGAN with text descriptions as inputs to generate images.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published