vittoriopippi commited on
Commit
9c772c4
Β·
1 Parent(s): 7248816

Change imports

Browse files
__init__.py ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ from .modeling_vatrpp import VATrPP
2
+ from .configuration_vatrpp import VATrPPConfig
data/dataset.py CHANGED
@@ -1,7 +1,6 @@
1
  import random
2
  from collections import defaultdict
3
 
4
- import matplotlib.pyplot as plt
5
  import torch
6
  from torch.utils.data import Dataset
7
  import torchvision.transforms as transforms
 
1
  import random
2
  from collections import defaultdict
3
 
 
4
  import torch
5
  from torch.utils.data import Dataset
6
  import torchvision.transforms as transforms
modeling_vatrpp.py CHANGED
@@ -6,9 +6,9 @@ import cv2
6
  import numpy as np
7
  import torch
8
 
9
- from data.dataset import FolderDataset
10
- from models.model import VATr
11
- from util.vision import detect_text_bounds
12
  from torchvision.transforms.functional import to_pil_image
13
 
14
 
 
6
  import numpy as np
7
  import torch
8
 
9
+ from .data.dataset import FolderDataset
10
+ from .models.model import VATr
11
+ from .models.util.vision import detect_text_bounds
12
  from torchvision.transforms.functional import to_pil_image
13
 
14
 
models/BigGAN_networks.py CHANGED
@@ -9,14 +9,14 @@ import torch.nn as nn
9
  import torch.nn.functional as F
10
  import random
11
 
12
- from util.augmentations import ProgressiveWordCrop, CycleWordCrop, StaticWordCrop, RandomWordCrop
13
  from . import BigGAN_layers as layers
14
  from .networks import init_weights
15
  import torchvision
16
  # Attention is passed in in the format '32_64' to mean applying an attention
17
  # block at both resolution 32x32 and 64x64. Just '64' will apply at 64x64.
18
 
19
- from models.blocks import Conv2dBlock, ResBlocks
20
 
21
 
22
  # Discriminator architecture, same paradigm as G's above
 
9
  import torch.nn.functional as F
10
  import random
11
 
12
+ from .util.augmentations import ProgressiveWordCrop, CycleWordCrop, StaticWordCrop, RandomWordCrop
13
  from . import BigGAN_layers as layers
14
  from .networks import init_weights
15
  import torchvision
16
  # Attention is passed in in the format '32_64' to mean applying an attention
17
  # block at both resolution 32x32 and 64x64. Just '64' will apply at 64x64.
18
 
19
+ from .blocks import Conv2dBlock, ResBlocks
20
 
21
 
22
  # Discriminator architecture, same paradigm as G's above
models/model.py CHANGED
@@ -4,19 +4,19 @@ from torch.nn.utils import clip_grad_norm_
4
  import sys
5
  import torchvision.models as models
6
 
7
- from models.inception import InceptionV3
8
- from models.transformer import *
9
- from util.augmentations import OCRAugment
10
- from util.misc import SmoothedValue
11
- from util.text import get_generator, AugmentedGenerator
12
  from .BigGAN_networks import *
13
  from .OCR_network import *
14
- from models.blocks import Conv2dBlock, ResBlocks
15
- from util.util import loss_hinge_dis, loss_hinge_gen, make_one_hot
16
 
17
- import models.config as config
18
  from .positional_encodings import PositionalEncoding1D
19
- from models.unifont_module import UnifontModule
20
  from PIL import Image
21
 
22
 
 
4
  import sys
5
  import torchvision.models as models
6
 
7
+ from .inception import InceptionV3
8
+ from .transformer import *
9
+ from .util.augmentations import OCRAugment
10
+ from .util.misc import SmoothedValue
11
+ from .util.text import get_generator, AugmentedGenerator
12
  from .BigGAN_networks import *
13
  from .OCR_network import *
14
+ from .blocks import Conv2dBlock, ResBlocks
15
+ from .util.util import loss_hinge_dis, loss_hinge_gen, make_one_hot
16
 
17
+ from . import config
18
  from .positional_encodings import PositionalEncoding1D
19
+ from .unifont_module import UnifontModule
20
  from PIL import Image
21
 
22
 
models/networks.py CHANGED
@@ -3,7 +3,7 @@ import torch.nn as nn
3
  from torch.nn import init
4
  import functools
5
  from torch.optim import lr_scheduler
6
- from util.util import to_device, load_network
7
 
8
  ###############################################################################
9
  # Helper Functions
 
3
  from torch.nn import init
4
  import functools
5
  from torch.optim import lr_scheduler
6
+ from .util.util import to_device, load_network
7
 
8
  ###############################################################################
9
  # Helper Functions
{util β†’ models/util}/__init__.py RENAMED
File without changes
{util β†’ models/util}/augmentations.py RENAMED
@@ -11,7 +11,7 @@ import torchvision.transforms
11
  import torchvision.transforms.functional as F
12
  from matplotlib import pyplot as plt
13
 
14
- from data.dataset import CollectionTextDataset, TextDataset
15
 
16
 
17
  def to_opencv(batch: torch.Tensor):
 
11
  import torchvision.transforms.functional as F
12
  from matplotlib import pyplot as plt
13
 
14
+ # from data.dataset import CollectionTextDataset, TextDataset
15
 
16
 
17
  def to_opencv(batch: torch.Tensor):
{util β†’ models/util}/loading.py RENAMED
File without changes
{util β†’ models/util}/misc.py RENAMED
File without changes
{util β†’ models/util}/text.py RENAMED
File without changes
{util β†’ models/util}/util.py RENAMED
File without changes
{util β†’ models/util}/vision.py RENAMED
File without changes