renyurui / controllable-person-synthesis

Human pose manipulation for fashion

  • Public
  • 3.6K runs
  • T4
  • GitHub
  • Paper
  • License

Input

*file
Preview
reference_image

Input reference image

*file

Text file containing 2D array of Openpose keypoints (shape 18x2)

integer
(minimum: -1, maximum: 4294967295)

Seed for random number generator. If -1, a random seed will be chosen. (minimum: -1; maximum: 4294967295)

Default: -1

Output

result

result

target_skeleton

target_skeleton
Generated in

Run time and cost

This model costs approximately $0.00042 to run on Replicate, or 2380 runs per $1, but this varies depending on your inputs. It is also open source and you can run it on your own computer with Docker.

This model runs on Nvidia T4 GPU hardware. Predictions typically complete within 2 seconds. The predict time for this model varies significantly based on the inputs.

Readme

ArXiv | Get Start

Neural-Texture-Extraction-Distribution

The PyTorch implementation for our paper “Neural Texture Extraction and Distribution for Controllable Person Image Synthesis” (CVPR2022 Oral)

We propose a Neural-Texture-Extraction-Distribution operation for controllable person image synthesis. Our model can be used to control the pose and appearance of a reference image:

NOTE: This demo only supports pose control. For appearance control, please see the original repo to run models locally.

  • Pose Control

Usage

The model takes in as input a reference image, as well a .txt file containing a 18x2 array of OpenPose keypoints. See Pose Output Format: COCO here.

Ensure that the reference image has a clean, monocolor background to optimize feature extraction and prevent artifacts. Also note that the model is trained on a limited range of demographics, which may cause potential artifacts.

The model outputs a visualization of the target skeleton, as well as the reference model manipulated to fit the target skeleton pose.