vittoriopippi
commited on
Commit
Β·
9c772c4
1
Parent(s):
7248816
Change imports
Browse files- __init__.py +2 -0
- data/dataset.py +0 -1
- modeling_vatrpp.py +3 -3
- models/BigGAN_networks.py +2 -2
- models/model.py +9 -9
- models/networks.py +1 -1
- {util β models/util}/__init__.py +0 -0
- {util β models/util}/augmentations.py +1 -1
- {util β models/util}/loading.py +0 -0
- {util β models/util}/misc.py +0 -0
- {util β models/util}/text.py +0 -0
- {util β models/util}/util.py +0 -0
- {util β models/util}/vision.py +0 -0
__init__.py
ADDED
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
1 |
+
from .modeling_vatrpp import VATrPP
|
2 |
+
from .configuration_vatrpp import VATrPPConfig
|
data/dataset.py
CHANGED
@@ -1,7 +1,6 @@
|
|
1 |
import random
|
2 |
from collections import defaultdict
|
3 |
|
4 |
-
import matplotlib.pyplot as plt
|
5 |
import torch
|
6 |
from torch.utils.data import Dataset
|
7 |
import torchvision.transforms as transforms
|
|
|
1 |
import random
|
2 |
from collections import defaultdict
|
3 |
|
|
|
4 |
import torch
|
5 |
from torch.utils.data import Dataset
|
6 |
import torchvision.transforms as transforms
|
modeling_vatrpp.py
CHANGED
@@ -6,9 +6,9 @@ import cv2
|
|
6 |
import numpy as np
|
7 |
import torch
|
8 |
|
9 |
-
from data.dataset import FolderDataset
|
10 |
-
from models.model import VATr
|
11 |
-
from util.vision import detect_text_bounds
|
12 |
from torchvision.transforms.functional import to_pil_image
|
13 |
|
14 |
|
|
|
6 |
import numpy as np
|
7 |
import torch
|
8 |
|
9 |
+
from .data.dataset import FolderDataset
|
10 |
+
from .models.model import VATr
|
11 |
+
from .models.util.vision import detect_text_bounds
|
12 |
from torchvision.transforms.functional import to_pil_image
|
13 |
|
14 |
|
models/BigGAN_networks.py
CHANGED
@@ -9,14 +9,14 @@ import torch.nn as nn
|
|
9 |
import torch.nn.functional as F
|
10 |
import random
|
11 |
|
12 |
-
from util.augmentations import ProgressiveWordCrop, CycleWordCrop, StaticWordCrop, RandomWordCrop
|
13 |
from . import BigGAN_layers as layers
|
14 |
from .networks import init_weights
|
15 |
import torchvision
|
16 |
# Attention is passed in in the format '32_64' to mean applying an attention
|
17 |
# block at both resolution 32x32 and 64x64. Just '64' will apply at 64x64.
|
18 |
|
19 |
-
from
|
20 |
|
21 |
|
22 |
# Discriminator architecture, same paradigm as G's above
|
|
|
9 |
import torch.nn.functional as F
|
10 |
import random
|
11 |
|
12 |
+
from .util.augmentations import ProgressiveWordCrop, CycleWordCrop, StaticWordCrop, RandomWordCrop
|
13 |
from . import BigGAN_layers as layers
|
14 |
from .networks import init_weights
|
15 |
import torchvision
|
16 |
# Attention is passed in in the format '32_64' to mean applying an attention
|
17 |
# block at both resolution 32x32 and 64x64. Just '64' will apply at 64x64.
|
18 |
|
19 |
+
from .blocks import Conv2dBlock, ResBlocks
|
20 |
|
21 |
|
22 |
# Discriminator architecture, same paradigm as G's above
|
models/model.py
CHANGED
@@ -4,19 +4,19 @@ from torch.nn.utils import clip_grad_norm_
|
|
4 |
import sys
|
5 |
import torchvision.models as models
|
6 |
|
7 |
-
from
|
8 |
-
from
|
9 |
-
from util.augmentations import OCRAugment
|
10 |
-
from util.misc import SmoothedValue
|
11 |
-
from util.text import get_generator, AugmentedGenerator
|
12 |
from .BigGAN_networks import *
|
13 |
from .OCR_network import *
|
14 |
-
from
|
15 |
-
from util.util import loss_hinge_dis, loss_hinge_gen, make_one_hot
|
16 |
|
17 |
-
|
18 |
from .positional_encodings import PositionalEncoding1D
|
19 |
-
from
|
20 |
from PIL import Image
|
21 |
|
22 |
|
|
|
4 |
import sys
|
5 |
import torchvision.models as models
|
6 |
|
7 |
+
from .inception import InceptionV3
|
8 |
+
from .transformer import *
|
9 |
+
from .util.augmentations import OCRAugment
|
10 |
+
from .util.misc import SmoothedValue
|
11 |
+
from .util.text import get_generator, AugmentedGenerator
|
12 |
from .BigGAN_networks import *
|
13 |
from .OCR_network import *
|
14 |
+
from .blocks import Conv2dBlock, ResBlocks
|
15 |
+
from .util.util import loss_hinge_dis, loss_hinge_gen, make_one_hot
|
16 |
|
17 |
+
from . import config
|
18 |
from .positional_encodings import PositionalEncoding1D
|
19 |
+
from .unifont_module import UnifontModule
|
20 |
from PIL import Image
|
21 |
|
22 |
|
models/networks.py
CHANGED
@@ -3,7 +3,7 @@ import torch.nn as nn
|
|
3 |
from torch.nn import init
|
4 |
import functools
|
5 |
from torch.optim import lr_scheduler
|
6 |
-
from util.util import to_device, load_network
|
7 |
|
8 |
###############################################################################
|
9 |
# Helper Functions
|
|
|
3 |
from torch.nn import init
|
4 |
import functools
|
5 |
from torch.optim import lr_scheduler
|
6 |
+
from .util.util import to_device, load_network
|
7 |
|
8 |
###############################################################################
|
9 |
# Helper Functions
|
{util β models/util}/__init__.py
RENAMED
File without changes
|
{util β models/util}/augmentations.py
RENAMED
@@ -11,7 +11,7 @@ import torchvision.transforms
|
|
11 |
import torchvision.transforms.functional as F
|
12 |
from matplotlib import pyplot as plt
|
13 |
|
14 |
-
from data.dataset import CollectionTextDataset, TextDataset
|
15 |
|
16 |
|
17 |
def to_opencv(batch: torch.Tensor):
|
|
|
11 |
import torchvision.transforms.functional as F
|
12 |
from matplotlib import pyplot as plt
|
13 |
|
14 |
+
# from data.dataset import CollectionTextDataset, TextDataset
|
15 |
|
16 |
|
17 |
def to_opencv(batch: torch.Tensor):
|
{util β models/util}/loading.py
RENAMED
File without changes
|
{util β models/util}/misc.py
RENAMED
File without changes
|
{util β models/util}/text.py
RENAMED
File without changes
|
{util β models/util}/util.py
RENAMED
File without changes
|
{util β models/util}/vision.py
RENAMED
File without changes
|