I've taken a few pre-trained models and made an interactive web thing for trying them out. Chrome is recommended. The pix2pix model works by training on pairs of images such as building facade labels to building facades, and then attempts to generate the corresponding output image from any input image you give it.
The idea is straight from the pix2pix paper , which is a good read. Trained on about 2k stock cat photos and edges automatically generated from those photos.
Generates cat-colored objects, some with nightmare faces. We assume that you have installed caffe and pycaffe in your system. If not, see the official website for installation instructions. Once caffe is successfully installed, download the pre-trained FCN-8s semantic segmentation model MB by running.
Then make sure. If not, run the following command to add it. Further notes : Our pre-trained FCN model is not supposed to work on Cityscapes in the original resolution x as it was trained on x images that are then upsampled to x during training.
The purpose of the resizing during training was to 1 keep the label maps in the original high resolution untouched and 2 avoid the need to change the standard FCN training code and the architecture for Cityscapes.
During test time, you need to synthesize x results. Our test code will automatically upsample your results to x before feeding them to the pre-trained FCN model.
The output is at x resolution and will be compared to x ground truth labels. You do not need to resize the ground truth labels. The best way to verify whether everything is correct is to reproduce the numbers for real images in the paper first. Optionally, for displaying images during training and test, use the display package.
By default, the server listens on localhost. Pass 0. L1 error is plotted to the display by default. If you love cats, and love reading cool graphics, vision, and learning papers, please check out the Cat Paper Collection: [Github] [Webpage]. In pix2pix, testing mode is still setup to take image pairs like in training mode, where there is an X and a Y. Play Pix2Pix , a free online drawing game provided by GamesButler. Pix2Pix is a fun game that can be played on any device.
Interactive Image Translation with pix2pix -tensorflow. Alternative Face Mario Klingemann used our code to translate the appearance of French singer Francoise Hardy onto Kellyanne Conway's infamous "alternative facts" interview. Person-to-Person Brannon Dorsey recorded himself mimicking frames from a video of Ray Kurzweil giving a talk. Background masking Kaihu Chen performed a number of interesting experiments using our method, including getting it to mask out the background of a portrait as shown above.
Color palette completion Colormind adapted our code to predict a complete 5-color palette given a subset of the palette as input. Recent Related Work Generative adversarial networks have been vigorously explored in the last two years, and many conditional variants have been proposed. Please see the discussion of related work in our paper.
Below we point out three papers that especially influenced this work: the original GAN paper from Goodfellow et al. Ian J. Generative Adversarial Networks. NIPS, ICLR, ECCV,
0コメント