Skip to content

Latest commit

 

History

History
120 lines (81 loc) · 4.29 KB

File metadata and controls

120 lines (81 loc) · 4.29 KB

Structure-emphasized Multimodal Style Transfer

DOI

Pytorch(1.0+) implementation of My master paper "Structure-emphasized Multimodal Style Transfer".

We proposed 2 models, called SEMST_Original and SEMST_Auto in this work. More details can be founed in the paper.

This repository provides pre-trained models for you to generate your own image given content image and style image. Also, you can download the training dataset or prepare your own dataset to train the model from scratch.

If you have any question, please feel free to contact me. (Language in English/Japanese/Chinese will be ok!)


If you find this work useful for you, please cite it as follow in your paper. Thanks a lot.

@misc{Chen2020,
  author = {Chen Chen},
  title = {Structure-emphasized Multimodal Style Transfer},
  year = {2020},
  month = 1,
  doi = 10.5281/zenodo.3602064
  publisher = {Zenodo},
  url = {https://doi.org/10.5281/zenodo.3602064},
}

Requirements

  • Python 3.7+
  • PyTorch 1.0+
  • TorchVision
  • Pillow

Anaconda environment recommended here!

(optional)

  • GPU environment

Result

Some results of content image will be shown here.

image image image image image image image image



Notice: The train and test procedures as follow are the same for SEMST_Original and SEMST_Auto.



Test

  1. Clone this repository

    git clone https://github.com/irasin/Structure-emphasized-Multimodal-Style-Transfer
    cd Structure-emphasized-Multimodal-Style-Transfer
    cd SEMST_XXX(XXX means Original or Auto)
  2. Prepare your content image and style image. I provide some in the content and style and you can try to use them easily.

  3. Download the pretrained model SEMST_Original, SEMST_Auto and put them under the SEMST_XXX respectively.

  4. Generate the output image. A transferred output image w/&w/o style image and a NST_demo_like image will be generated.

    python test.py -c content_image_path -s style_image_path
usage: test.py [-h] [--content CONTENT] [--style STYLE]
              [--output_name OUTPUT_NAME] [--alpha ALPHA] [--gpu GPU]
              [--model_state_path MODEL_STATE_PATH]

If output_name is not given, it will use the combination of content image name and style image name.


Train

  1. Download COCO (as content dataset)and Wikiart (as style dataset) and unzip them, rename them as content and style respectively (recommended).

  2. Modify the argument in the train.py such as the path of directory, epoch, learning_rate or you can add your own training code.

  3. Train the model using gpu.

  4. python train.py
    usage: train.py [-h] [--batch_size BATCH_SIZE] [--epoch EPOCH] [--gpu GPU]
                 [--learning_rate LEARNING_RATE]
                 [--snapshot_interval SNAPSHOT_INTERVAL] [--alpha ALPHA]
                 [--gamma GAMMA] [--train_content_dir TRAIN_CONTENT_DIR]
                 [--train_style_dir TRAIN_STYLE_DIR]
                 [--test_content_dir TEST_CONTENT_DIR]
                 [--test_style_dir TEST_STYLE_DIR] [--save_dir SAVE_DIR]
                 [--reuse REUSE]