Spaces:
Running
Running
Update transcription.py
Browse files- __pycache__/process.cpython-310.pyc +0 -0
- __pycache__/transcription.cpython-310.pyc +0 -0
- app.py +6 -4
- process.py +1 -9
- transcription.py +18 -6
__pycache__/process.cpython-310.pyc
CHANGED
Binary files a/__pycache__/process.cpython-310.pyc and b/__pycache__/process.cpython-310.pyc differ
|
|
__pycache__/transcription.cpython-310.pyc
CHANGED
Binary files a/__pycache__/transcription.cpython-310.pyc and b/__pycache__/transcription.cpython-310.pyc differ
|
|
app.py
CHANGED
@@ -13,13 +13,13 @@ transcripter = TranscriptionMaker()
|
|
13 |
app = Flask(__name__)
|
14 |
|
15 |
users = []
|
16 |
-
segments_dir = ""
|
17 |
transcription_text=""
|
18 |
harassment_keywords = [
|
19 |
"バカ", "馬鹿", "アホ", "死ね", "クソ", "うざい",
|
20 |
"きもい", "キモい", "ブス", "デブ", "ハゲ",
|
21 |
"セクハラ", "パワハラ", "モラハラ"
|
22 |
]
|
|
|
23 |
|
24 |
# トップページ(テンプレート: index.html)
|
25 |
@app.route('/')
|
@@ -62,10 +62,10 @@ def reset_member():
|
|
62 |
# 書き起こし作成エンドポイント
|
63 |
@app.route('/transcription',methods =['GET','POST'])
|
64 |
def transcription():
|
65 |
-
global segments_dir
|
66 |
global transcription_text
|
|
|
67 |
try:
|
68 |
-
transcription_text = transcripter.create_transcription(
|
69 |
print(transcription_text)
|
70 |
with open(transcription_text,'r',encoding='utf-8') as file:
|
71 |
file_content = file.read()
|
@@ -109,7 +109,7 @@ def analyze():
|
|
109 |
# 音声アップロード&解析エンドポイント
|
110 |
@app.route('/upload_audio', methods=['POST'])
|
111 |
def upload_audio():
|
112 |
-
global
|
113 |
try:
|
114 |
data = request.get_json()
|
115 |
# name か users のいずれかが必須。どちらも無い場合はエラー
|
@@ -139,12 +139,14 @@ def upload_audio():
|
|
139 |
if len(users) > 1:
|
140 |
print("複数人の場合の処理")
|
141 |
matched_times, segments_dir = process.process_multi_audio(reference_paths, audio_path, threshold=0.05)
|
|
|
142 |
# 各メンバーのrateを計算
|
143 |
total_time = sum(matched_times)
|
144 |
rates = [(time / total_time) * 100 if total_time > 0 else 0 for time in matched_times]
|
145 |
return jsonify({"rates": rates}), 200
|
146 |
else:
|
147 |
matched_time, unmatched_time, segments_dir = process.process_audio(reference_paths[0], audio_path, threshold=0.05)
|
|
|
148 |
total_time = matched_time + unmatched_time
|
149 |
rate = (matched_time / total_time) * 100 if total_time > 0 else 0
|
150 |
return jsonify({"rate": rate}), 200
|
|
|
13 |
app = Flask(__name__)
|
14 |
|
15 |
users = []
|
|
|
16 |
transcription_text=""
|
17 |
harassment_keywords = [
|
18 |
"バカ", "馬鹿", "アホ", "死ね", "クソ", "うざい",
|
19 |
"きもい", "キモい", "ブス", "デブ", "ハゲ",
|
20 |
"セクハラ", "パワハラ", "モラハラ"
|
21 |
]
|
22 |
+
total_audio = ""
|
23 |
|
24 |
# トップページ(テンプレート: index.html)
|
25 |
@app.route('/')
|
|
|
62 |
# 書き起こし作成エンドポイント
|
63 |
@app.route('/transcription',methods =['GET','POST'])
|
64 |
def transcription():
|
|
|
65 |
global transcription_text
|
66 |
+
global total_audio
|
67 |
try:
|
68 |
+
transcription_text = transcripter.create_transcription(total_audio)
|
69 |
print(transcription_text)
|
70 |
with open(transcription_text,'r',encoding='utf-8') as file:
|
71 |
file_content = file.read()
|
|
|
109 |
# 音声アップロード&解析エンドポイント
|
110 |
@app.route('/upload_audio', methods=['POST'])
|
111 |
def upload_audio():
|
112 |
+
global total_audio
|
113 |
try:
|
114 |
data = request.get_json()
|
115 |
# name か users のいずれかが必須。どちらも無い場合はエラー
|
|
|
139 |
if len(users) > 1:
|
140 |
print("複数人の場合の処理")
|
141 |
matched_times, segments_dir = process.process_multi_audio(reference_paths, audio_path, threshold=0.05)
|
142 |
+
total_audio = transcripter.merge_segments(segments_dir)
|
143 |
# 各メンバーのrateを計算
|
144 |
total_time = sum(matched_times)
|
145 |
rates = [(time / total_time) * 100 if total_time > 0 else 0 for time in matched_times]
|
146 |
return jsonify({"rates": rates}), 200
|
147 |
else:
|
148 |
matched_time, unmatched_time, segments_dir = process.process_audio(reference_paths[0], audio_path, threshold=0.05)
|
149 |
+
total_audio = transcripter.merge_segments(segments_dir)
|
150 |
total_time = matched_time + unmatched_time
|
151 |
rate = (matched_time / total_time) * 100 if total_time > 0 else 0
|
152 |
return jsonify({"rate": rate}), 200
|
process.py
CHANGED
@@ -58,15 +58,7 @@ class AudioProcessor():
|
|
58 |
embedding2 = self.inference(path2)
|
59 |
return float(self.cosine_similarity(embedding1.data.flatten(), embedding2.data.flatten()))
|
60 |
|
61 |
-
|
62 |
-
letters = string.ascii_letters + string.digits
|
63 |
-
return ''.join(random.choice(letters) for i in range(length))
|
64 |
-
|
65 |
-
def generate_filename(self,random_length):
|
66 |
-
random_string = self.generate_random_string(random_length)
|
67 |
-
current_time = datetime.now().strftime("%Y%m%d%H%M%S")
|
68 |
-
filename = f"{current_time}_{random_string}.wav"
|
69 |
-
return filename
|
70 |
|
71 |
def process_audio(self, reference_path, input_path, output_folder='/tmp/data/matched_segments', seg_duration=1.0, threshold=0.5):
|
72 |
# 出力先ディレクトリの中身をクリアする
|
|
|
58 |
embedding2 = self.inference(path2)
|
59 |
return float(self.cosine_similarity(embedding1.data.flatten(), embedding2.data.flatten()))
|
60 |
|
61 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
62 |
|
63 |
def process_audio(self, reference_path, input_path, output_folder='/tmp/data/matched_segments', seg_duration=1.0, threshold=0.5):
|
64 |
# 出力先ディレクトリの中身をクリアする
|
transcription.py
CHANGED
@@ -1,6 +1,9 @@
|
|
1 |
import os
|
2 |
from faster_whisper import WhisperModel
|
3 |
from pydub import AudioSegment
|
|
|
|
|
|
|
4 |
|
5 |
# Matplotlibのキャッシュディレクトリを変更
|
6 |
os.environ["MPLCONFIGDIR"] = "/tmp/matplotlib"
|
@@ -18,10 +21,9 @@ class TranscriptionMaker():
|
|
18 |
|
19 |
|
20 |
#音声ファイルのディレクトリを受け取り、書き起こしファイルを作成する
|
21 |
-
def create_transcription(self,
|
22 |
results = []
|
23 |
-
|
24 |
-
audio_directory = self.merge_segments(segments_directory)
|
25 |
#ディレクトリ内のファイルを全て取得
|
26 |
if not os.path.isdir(audio_directory):
|
27 |
raise ValueError(f"The specified path is not a valid directory: {audio_directory}")
|
@@ -53,7 +55,7 @@ class TranscriptionMaker():
|
|
53 |
return output_file
|
54 |
|
55 |
#ファイル名が連続しているならくっつける
|
56 |
-
def merge_segments(self,segments_dir,output_dir = "/tmp/data/
|
57 |
if not os.path.exists(output_dir):
|
58 |
os.makedirs(output_dir, exist_ok=True)
|
59 |
|
@@ -90,7 +92,17 @@ class TranscriptionMaker():
|
|
90 |
segment = AudioSegment.from_file(file_path)
|
91 |
combined_audio += segment
|
92 |
# 出力ファイル名を設定して保存
|
93 |
-
output_file = os.path.join(output_dir,
|
94 |
combined_audio.export(output_file, format='wav')
|
95 |
|
96 |
-
return output_dir
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
import os
|
2 |
from faster_whisper import WhisperModel
|
3 |
from pydub import AudioSegment
|
4 |
+
import string
|
5 |
+
import random
|
6 |
+
from datetime import datetime
|
7 |
|
8 |
# Matplotlibのキャッシュディレクトリを変更
|
9 |
os.environ["MPLCONFIGDIR"] = "/tmp/matplotlib"
|
|
|
21 |
|
22 |
|
23 |
#音声ファイルのディレクトリを受け取り、書き起こしファイルを作成する
|
24 |
+
def create_transcription(self,audio_directory):
|
25 |
results = []
|
26 |
+
|
|
|
27 |
#ディレクトリ内のファイルを全て取得
|
28 |
if not os.path.isdir(audio_directory):
|
29 |
raise ValueError(f"The specified path is not a valid directory: {audio_directory}")
|
|
|
55 |
return output_file
|
56 |
|
57 |
#ファイル名が連続しているならくっつける
|
58 |
+
def merge_segments(self,segments_dir,output_dir = "/tmp/data/merged_segment"):
|
59 |
if not os.path.exists(output_dir):
|
60 |
os.makedirs(output_dir, exist_ok=True)
|
61 |
|
|
|
92 |
segment = AudioSegment.from_file(file_path)
|
93 |
combined_audio += segment
|
94 |
# 出力ファイル名を設定して保存
|
95 |
+
output_file = os.path.join(output_dir, self.generate_filename(3))
|
96 |
combined_audio.export(output_file, format='wav')
|
97 |
|
98 |
+
return output_dir
|
99 |
+
|
100 |
+
def generate_random_string(self,length):
|
101 |
+
letters = string.ascii_letters + string.digits
|
102 |
+
return ''.join(random.choice(letters) for i in range(length))
|
103 |
+
|
104 |
+
def generate_filename(self,random_length):
|
105 |
+
random_string = self.generate_random_string(random_length)
|
106 |
+
current_time = datetime.now().strftime("%Y%m%d%H%M%S")
|
107 |
+
filename = f"{current_time}_{random_string}.wav"
|
108 |
+
return filename
|