Spaces:
Configuration error
Configuration error
Fedir Zadniprovskyi
commited on
Commit
·
5741d7c
1
Parent(s):
d7d34e3
fix: circular import
Browse files- Taskfile.yaml +1 -1
- speaches/config.py +9 -2
- speaches/main.py +3 -2
- speaches/server_models.py +0 -11
Taskfile.yaml
CHANGED
@@ -15,7 +15,7 @@ tasks:
|
|
15 |
create-multi-arch-builder: docker buildx create --name main --driver=docker-container
|
16 |
build-and-push:
|
17 |
cmds:
|
18 |
-
- docker compose build --builder main
|
19 |
sources:
|
20 |
- Dockerfile.*
|
21 |
- speaches/*.py
|
|
|
15 |
create-multi-arch-builder: docker buildx create --name main --driver=docker-container
|
16 |
build-and-push:
|
17 |
cmds:
|
18 |
+
- docker compose build --builder main --push
|
19 |
sources:
|
20 |
- Dockerfile.*
|
21 |
- speaches/*.py
|
speaches/config.py
CHANGED
@@ -3,8 +3,6 @@ import enum
|
|
3 |
from pydantic import BaseModel, Field
|
4 |
from pydantic_settings import BaseSettings, SettingsConfigDict
|
5 |
|
6 |
-
from speaches.server_models import ResponseFormat
|
7 |
-
|
8 |
SAMPLES_PER_SECOND = 16000
|
9 |
BYTES_PER_SAMPLE = 2
|
10 |
BYTES_PER_SECOND = SAMPLES_PER_SECOND * BYTES_PER_SAMPLE
|
@@ -12,6 +10,15 @@ BYTES_PER_SECOND = SAMPLES_PER_SECOND * BYTES_PER_SAMPLE
|
|
12 |
# 1 SECOND OF AUDIO = 32000 BYTES = 16000 SAMPLES
|
13 |
|
14 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
# https://huggingface.co/Systran
|
16 |
class Model(enum.StrEnum):
|
17 |
TINY_EN = "tiny.en"
|
|
|
3 |
from pydantic import BaseModel, Field
|
4 |
from pydantic_settings import BaseSettings, SettingsConfigDict
|
5 |
|
|
|
|
|
6 |
SAMPLES_PER_SECOND = 16000
|
7 |
BYTES_PER_SAMPLE = 2
|
8 |
BYTES_PER_SECOND = SAMPLES_PER_SECOND * BYTES_PER_SAMPLE
|
|
|
10 |
# 1 SECOND OF AUDIO = 32000 BYTES = 16000 SAMPLES
|
11 |
|
12 |
|
13 |
+
# https://platform.openai.com/docs/api-reference/audio/createTranscription#audio-createtranscription-response_format
|
14 |
+
class ResponseFormat(enum.StrEnum):
|
15 |
+
TEXT = "text"
|
16 |
+
JSON = "json"
|
17 |
+
VERBOSE_JSON = "verbose_json"
|
18 |
+
# VTT = "vtt"
|
19 |
+
# SRT = "srt"
|
20 |
+
|
21 |
+
|
22 |
# https://huggingface.co/Systran
|
23 |
class Model(enum.StrEnum):
|
24 |
TINY_EN = "tiny.en"
|
speaches/main.py
CHANGED
@@ -17,10 +17,11 @@ from faster_whisper.vad import VadOptions, get_speech_timestamps
|
|
17 |
from speaches import utils
|
18 |
from speaches.asr import FasterWhisperASR
|
19 |
from speaches.audio import AudioStream, audio_samples_from_file
|
20 |
-
from speaches.config import SAMPLES_PER_SECOND, Language, Model,
|
|
|
21 |
from speaches.core import Transcription
|
22 |
from speaches.logger import logger
|
23 |
-
from speaches.server_models import (
|
24 |
TranscriptionVerboseJsonResponse)
|
25 |
from speaches.transcriber import audio_transcriber
|
26 |
|
|
|
17 |
from speaches import utils
|
18 |
from speaches.asr import FasterWhisperASR
|
19 |
from speaches.audio import AudioStream, audio_samples_from_file
|
20 |
+
from speaches.config import (SAMPLES_PER_SECOND, Language, Model,
|
21 |
+
ResponseFormat, config)
|
22 |
from speaches.core import Transcription
|
23 |
from speaches.logger import logger
|
24 |
+
from speaches.server_models import (TranscriptionJsonResponse,
|
25 |
TranscriptionVerboseJsonResponse)
|
26 |
from speaches.transcriber import audio_transcriber
|
27 |
|
speaches/server_models.py
CHANGED
@@ -1,7 +1,5 @@
|
|
1 |
from __future__ import annotations
|
2 |
|
3 |
-
import enum
|
4 |
-
|
5 |
from faster_whisper.transcribe import Segment, TranscriptionInfo, Word
|
6 |
from pydantic import BaseModel
|
7 |
|
@@ -9,15 +7,6 @@ from speaches import utils
|
|
9 |
from speaches.core import Transcription
|
10 |
|
11 |
|
12 |
-
# https://platform.openai.com/docs/api-reference/audio/createTranscription#audio-createtranscription-response_format
|
13 |
-
class ResponseFormat(enum.StrEnum):
|
14 |
-
TEXT = "text"
|
15 |
-
JSON = "json"
|
16 |
-
VERBOSE_JSON = "verbose_json"
|
17 |
-
# VTT = "vtt"
|
18 |
-
# SRT = "srt"
|
19 |
-
|
20 |
-
|
21 |
# https://platform.openai.com/docs/api-reference/audio/json-object
|
22 |
class TranscriptionJsonResponse(BaseModel):
|
23 |
text: str
|
|
|
1 |
from __future__ import annotations
|
2 |
|
|
|
|
|
3 |
from faster_whisper.transcribe import Segment, TranscriptionInfo, Word
|
4 |
from pydantic import BaseModel
|
5 |
|
|
|
7 |
from speaches.core import Transcription
|
8 |
|
9 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
# https://platform.openai.com/docs/api-reference/audio/json-object
|
11 |
class TranscriptionJsonResponse(BaseModel):
|
12 |
text: str
|