Jie Hu commited on
Commit
d17054d
·
1 Parent(s): 61e5f12

init project

Browse files
Files changed (1) hide show
  1. app.py +7 -3
app.py CHANGED
@@ -37,10 +37,14 @@ from modules.mobilesamv2.utils.transforms import ResizeLongestSide
37
  from modules.pe3r.models import Models
38
  import torchvision.transforms as tvf
39
 
 
40
 
41
  silent = False
42
  device = 'cuda' if torch.cuda.is_available() else 'cpu'
43
- pe3r = Models(device)
 
 
 
44
 
45
 
46
  def _convert_scene_output_to_glb(outdir, imgs, pts3d, mask, focals, cams2world, cam_size=0.05,
@@ -469,7 +473,7 @@ def get_reconstructed_scene(outdir, filelist, schedule, niter, min_conf_thr,
469
  scenegraph_type = scenegraph_type + "-" + str(refid)
470
 
471
  pairs = make_pairs(imgs, scene_graph=scenegraph_type, prefilter=None, symmetrize=True)
472
- output = inference(pairs, pe3r.mast3r, device, batch_size=1, verbose=not silent)
473
  mode = GlobalAlignerMode.PointCloudOptimizer if len(imgs) > 2 else GlobalAlignerMode.PairViewer
474
  scene_1 = global_aligner(output, cog_seg_maps, rev_cog_seg_maps, cog_feats, device=device, mode=mode, verbose=not silent)
475
  lr = 0.01
@@ -482,7 +486,7 @@ def get_reconstructed_scene(outdir, filelist, schedule, niter, min_conf_thr,
482
  # print(imgs[i]['img'].shape, scene.imgs[i].shape, ImgNorm(scene.imgs[i])[None])
483
  imgs[i]['img'] = ImgNorm(scene_1.imgs[i])[None]
484
  pairs = make_pairs(imgs, scene_graph=scenegraph_type, prefilter=None, symmetrize=True)
485
- output = inference(pairs, pe3r.mast3r, device, batch_size=1, verbose=not silent)
486
  mode = GlobalAlignerMode.PointCloudOptimizer if len(imgs) > 2 else GlobalAlignerMode.PairViewer
487
  scene = global_aligner(output, cog_seg_maps, rev_cog_seg_maps, cog_feats, device=device, mode=mode, verbose=not silent)
488
  ori_imgs = scene.ori_imgs
 
37
  from modules.pe3r.models import Models
38
  import torchvision.transforms as tvf
39
 
40
+ from modules.mast3r.model import AsymmetricMASt3R
41
 
42
  silent = False
43
  device = 'cuda' if torch.cuda.is_available() else 'cpu'
44
+ # pe3r = Models(device)
45
+ MAST3R_CKP = 'naver/MASt3R_ViTLarge_BaseDecoder_512_catmlpdpt_metric'
46
+ mast3r = AsymmetricMASt3R.from_pretrained(MAST3R_CKP).to(device)
47
+
48
 
49
 
50
  def _convert_scene_output_to_glb(outdir, imgs, pts3d, mask, focals, cams2world, cam_size=0.05,
 
473
  scenegraph_type = scenegraph_type + "-" + str(refid)
474
 
475
  pairs = make_pairs(imgs, scene_graph=scenegraph_type, prefilter=None, symmetrize=True)
476
+ output = inference(pairs, mast3r, device, batch_size=1, verbose=not silent)
477
  mode = GlobalAlignerMode.PointCloudOptimizer if len(imgs) > 2 else GlobalAlignerMode.PairViewer
478
  scene_1 = global_aligner(output, cog_seg_maps, rev_cog_seg_maps, cog_feats, device=device, mode=mode, verbose=not silent)
479
  lr = 0.01
 
486
  # print(imgs[i]['img'].shape, scene.imgs[i].shape, ImgNorm(scene.imgs[i])[None])
487
  imgs[i]['img'] = ImgNorm(scene_1.imgs[i])[None]
488
  pairs = make_pairs(imgs, scene_graph=scenegraph_type, prefilter=None, symmetrize=True)
489
+ output = inference(pairs, mast3r, device, batch_size=1, verbose=not silent)
490
  mode = GlobalAlignerMode.PointCloudOptimizer if len(imgs) > 2 else GlobalAlignerMode.PairViewer
491
  scene = global_aligner(output, cog_seg_maps, rev_cog_seg_maps, cog_feats, device=device, mode=mode, verbose=not silent)
492
  ori_imgs = scene.ori_imgs