Niral Patel commited on
Commit
fee470d
·
1 Parent(s): f17f988

change in model config

Browse files
.gitattributes CHANGED
@@ -36,3 +36,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
36
  *.tar.gz filter=lfs diff=lfs merge=lfs -text
37
  model.data-00000-of-00001 filter=lfs diff=lfs merge=lfs -text
38
  vocals.wav filter=lfs diff=lfs merge=lfs -text
 
 
36
  *.tar.gz filter=lfs diff=lfs merge=lfs -text
37
  model.data-00000-of-00001 filter=lfs diff=lfs merge=lfs -text
38
  vocals.wav filter=lfs diff=lfs merge=lfs -text
39
+ pretrained_models/2stems/model.data-00000-of-00001 filter=lfs diff=lfs merge=lfs -text
.gitignore CHANGED
@@ -1 +1,2 @@
1
- env/
 
 
1
+ env/
2
+ __pycache__/
custom_model.py CHANGED
@@ -25,7 +25,7 @@ class SpleeterModel(PreTrainedModel):
25
  Returns:
26
  dict: Separated stems.
27
  """
28
- return self.separator.separate(audio_path)
29
 
30
 
31
  AutoConfig.register("spleeter", SpleeterConfig)
 
25
  Returns:
26
  dict: Separated stems.
27
  """
28
+ return self.separator.separate_to_file(audio_path, "separated_audio")
29
 
30
 
31
  AutoConfig.register("spleeter", SpleeterConfig)
pretrained_models/2stems/._checkpoint ADDED
Binary file (176 Bytes). View file
 
pretrained_models/2stems/.probe ADDED
@@ -0,0 +1 @@
 
 
1
+ OK
pretrained_models/2stems/checkpoint ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ model_checkpoint_path: "model"
2
+ all_model_checkpoint_paths: "model"
pretrained_models/2stems/model.data-00000-of-00001 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7747f9fd2c782306dbec1504360fbb645a097a48446f23486c3ff9c89bc11788
3
+ size 78614080
pretrained_models/2stems/model.index ADDED
Binary file (5.24 kB). View file
 
pretrained_models/2stems/model.meta ADDED
Binary file (806 kB). View file
 
test.py CHANGED
@@ -1,21 +1,6 @@
1
- # from transformers import AutoConfig, AutoModel
2
-
3
- # config = AutoConfig.from_pretrained("niral-env/youtube_spleeter")
4
- # print(config)
5
- # model = AutoModel.from_pretrained("niral-env/youtube_spleeter", config=config)
6
- # print(model)
7
-
8
-
9
- # Load model directly
10
- # from transformers import AutoModel
11
- # model = AutoModel.from_pretrained("niral-env/youtube_spleeter")
12
-
13
- # print(model)
14
-
15
  from transformers import AutoConfig, AutoModel
16
  from custom_model import SpleeterModel
17
 
18
- # Load configuration and model
19
  config = AutoConfig.from_pretrained("niral-env/youtube_spleeter")
20
  print("----"*30)
21
  print(config)
@@ -23,12 +8,15 @@ print("----"*30)
23
 
24
  model = SpleeterModel(config)
25
 
26
- # Process an audio file
27
- # result = model.forward("example.wav")
28
-
29
- # Output separated stems
30
  print(model)
31
  result = model.forward("vocals.wav")
32
 
33
- # Output separated stems
34
- print(result)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  from transformers import AutoConfig, AutoModel
2
  from custom_model import SpleeterModel
3
 
 
4
  config = AutoConfig.from_pretrained("niral-env/youtube_spleeter")
5
  print("----"*30)
6
  print(config)
 
8
 
9
  model = SpleeterModel(config)
10
 
 
 
 
 
11
  print(model)
12
  result = model.forward("vocals.wav")
13
 
14
+ print(result)
15
+
16
+
17
+ # from transformers import AutoModel
18
+ # model = AutoModel.from_pretrained("niral-env/youtube_spleeter")
19
+ # print(model)
20
+ # result = model.forward("vocals.wav")
21
+
22
+ # print(result)