|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import torch.nn as nn |
|
|
|
from huggan.pytorch.huggan_mixin import HugGANModelHubMixin |
|
|
|
|
|
class Generator(nn.Module, HugGANModelHubMixin): |
|
def __init__(self, num_channels=3, latent_dim=100, hidden_size=64): |
|
super(Generator, self).__init__() |
|
self.model = nn.Sequential( |
|
|
|
nn.ConvTranspose2d(latent_dim, hidden_size * 8, 4, 1, 0, bias=False), |
|
nn.BatchNorm2d(hidden_size * 8), |
|
nn.ReLU(True), |
|
|
|
nn.ConvTranspose2d(hidden_size * 8, hidden_size * 4, 4, 2, 1, bias=False), |
|
nn.BatchNorm2d(hidden_size * 4), |
|
nn.ReLU(True), |
|
|
|
nn.ConvTranspose2d(hidden_size * 4, hidden_size * 2, 4, 2, 1, bias=False), |
|
nn.BatchNorm2d(hidden_size * 2), |
|
nn.ReLU(True), |
|
|
|
nn.ConvTranspose2d(hidden_size * 2, hidden_size, 4, 2, 1, bias=False), |
|
nn.BatchNorm2d(hidden_size), |
|
nn.ReLU(True), |
|
|
|
nn.ConvTranspose2d(hidden_size, num_channels, 4, 2, 1, bias=False), |
|
nn.Tanh() |
|
|
|
) |
|
|
|
def forward(self, noise): |
|
pixel_values = self.model(noise) |
|
|
|
return pixel_values |
|
|
|
|
|
class Discriminator(nn.Module): |
|
def __init__(self, num_channels=3, hidden_size=64): |
|
super(Discriminator, self).__init__() |
|
self.model = nn.Sequential( |
|
|
|
nn.Conv2d(num_channels, hidden_size, 4, 2, 1, bias=False), |
|
nn.LeakyReLU(0.2, inplace=True), |
|
|
|
nn.Conv2d(hidden_size, hidden_size * 2, 4, 2, 1, bias=False), |
|
nn.BatchNorm2d(hidden_size * 2), |
|
nn.LeakyReLU(0.2, inplace=True), |
|
|
|
nn.Conv2d(hidden_size * 2, hidden_size * 4, 4, 2, 1, bias=False), |
|
nn.BatchNorm2d(hidden_size * 4), |
|
nn.LeakyReLU(0.2, inplace=True), |
|
|
|
nn.Conv2d(hidden_size * 4, hidden_size * 8, 4, 2, 1, bias=False), |
|
nn.BatchNorm2d(hidden_size * 8), |
|
nn.LeakyReLU(0.2, inplace=True), |
|
|
|
nn.Conv2d(hidden_size * 8, 1, 4, 1, 0, bias=False), |
|
nn.Sigmoid(), |
|
) |
|
|
|
def forward(self, pixel_values): |
|
logits = self.model(pixel_values) |
|
|
|
return logits |
|
|