Neural Style and MSG-Net in PyTorch

This repo provides PyTorch Implementation of MSG-Net (ours) and Neural Style (Gatys et al. CVPR 2016), which has been included by ModelDepot. We also provide Torch implementation and MXNet implementation. Tabe of content MSG-Net Multi-style Generative Network for Real-time Transfer [arXiv] [project] Hang Zhang, Kristin Dana @article{zhang2017multistyle, title={Multi-style Generative Network for Real-time Transfer}, author={Zhang, Hang and Dana, Kristin}, journal={arXiv preprint arXiv:1703.06953}, year={2017} } Stylize Images Using Pre-trained MSG-Net Download the pre-trained model

Read more

DeepLab resnet v2 model implementation in pytorch

DeepLab resnet v2 model implementation in pytorch. The architecture of deepLab-ResNet has been replicated exactly as it is from the caffe implementation. This architecture calculates losses on input images over multiple scales ( 1x, 0.75x, 0.5x ). Losses are calculated individually over these 3 scales. In addition to these 3 losses, one more loss is calculated after merging the output score maps on the 3 scales. These 4 losses are added to calculate the total loss. Updates 18 July 2017 […]

Read more

Attention is all you need: A Pytorch Implementation

This is a PyTorch implementation of the Transformer model in “Attention is All You Need” (Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin, arxiv, 2017). A novel sequence to sequence framework utilizes the self-attention mechanism, instead of Convolution operation or Recurrent structure, and achieve the state-of-the-art performance on WMT 2014 English-to-German translation task. (2017/06/12) The official Tensorflow Implementation can be found in: tensorflow/tensor2tensor. To learn more about self-attention mechanism, you could […]

Read more

A complete suite for training sequence-to-sequence models in PyTorch

This is a complete suite for training sequence-to-sequence models in PyTorch. It consists of several models and code to both train and infer using them. Using this code you can train: Neural-machine-translation (NMT) models Language models Image to caption generation Skip-thought sentence representations And more… Installation git clone –recursive https://github.com/eladhoffer/seq2seq.pytorch cd seq2seq.pytorch; python setup.py develop Models Models currently available: Datasets Datasets currently available: All datasets can be tokenized using 3 available segmentation methods: Character based segmentation Word based segmentation Byte-pair-encoding […]

Read more

BEGAN in PyTorch

This project is still in progress. If you are looking for the working code, use BEGAN-tensorflow. Requirements Usage First download CelebA datasets with: $ apt-get install p7zip-full # ubuntu $ brew install p7zip # Mac $ python download.py or you can use your own dataset by placing images like: data └── YOUR_DATASET_NAME ├── xxx.jpg (name doesn’t matter) ├── yyy.jpg └── … To train a model:

Read more

A PyTorch implementation of Multi-digit Number Recognition from Street View Imagery using Deep Convolutional Neural Networks

A PyTorch implementation of Multi-digit Number Recognition from Street View Imagery using Deep Convolutional Neural Networks If you’re interested in C++ inference, move HERE Results Steps GPU Batch Size Learning Rate Patience Decay Step Decay Rate Training Speed (FPS) Accuracy 54000 GTX 1080 Ti 512 0.16 100 625 0.9 ~1700 95.65% Sample $ python infer.py -c=./logs/model-54000.pth ./images/test-75.png length: 2 digits: 7 5 10 10 10

Read more

Wasserstein GAN with PyTorch

Code accompanying the paper “Wasserstein GAN” A few notes The first time running on the LSUN dataset it can take a long time (up to an hour) to create the dataloader. After the first run a small cache file will be created and the process should take a matter of seconds. The cache is a list of indices in the lmdb database (of LSUN) The only addition to the code (that we forgot, and will add, on the paper) are […]

Read more

PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation

This repo is implementation for PointNet(https://arxiv.org/abs/1612.00593) in pytorch. The model is in pointnet/model.py. It is tested with pytorch-1.0. git clone https://github.com/fxia22/pointnet.pytorch cd pointnet.pytorch pip install -e . Download and build visualization tool cd script bash build.sh #build C++ code for visualization bash download.sh #download dataset Training cd utils python train_classification.py –dataset –nepoch= –dataset_type python train_segmentation.py –dataset –nepoch= Use –feature_transform to use feature transform. Classification performance On ModelNet40: Overall Acc Original implementation 89.2 this implementation(w/o feature transform) 86.4 this implementation(w/ feature […]

Read more

Exotic structured image classifier in python

This implements one of result networks from Large-scale evolution of image classifiers by Esteban Real, et. al. Requirements Training Copy two files to {torchvision_path}/models cp {__init__.py,evloution.py} {torchvision_path}/models Run python main.py -a evolution {cifar10_data_dir} cf. How to know {torchvision_path}? import torchvision print(torchvision.__file__) Note The numbers of channels are not in the paper and it is set by me similar with vgg. You need to adjust these for better performance. If you want to adsjust    

Read more

The Django Template Language: Tags and Filters

Django is a powerful framework for creating web applications in Python. Its features include database models, routing URLs, authentication, user management, administrative tools, and a template language. You can compose reusable HTML that changes based on the data you pass to the template language. Django templates use tags and filters to define a mini-language that’s similar to Python—but isn’t Python. You’ll get to know Django templates through the tags and filters you use to compose reusable HTML. In this tutorial, […]

Read more
1 13 14 15 16 17 48