File size: 2,908 Bytes
49d6945
08192cc
81c3e39
49d6945
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5a962ae
81c3e39
49d6945
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
import tensorflow as tf
#from transformers import pipeline
from huggingface_hub import from_pretrained_keras
import pandas as pd
import numpy as np
import joblib
import os
import sys

# librosa is a Python library for analyzing audio and music. It can be used to extract the data from the audio files we will see it later.
import librosa
import librosa.display
import seaborn as sns
import matplotlib.pyplot as plt

from sklearn.preprocessing import StandardScaler, OneHotEncoder
from sklearn.metrics import confusion_matrix, classification_report
from sklearn.model_selection import train_test_split

# to play the audio files


import keras
from keras.preprocessing import sequence
from keras.models import Sequential, model_from_json
from keras.layers import Dense, Embedding
from keras.layers import LSTM, BatchNormalization, GRU
from keras.preprocessing.text import Tokenizer

from tensorflow.keras.utils import to_categorical
from keras.layers import Input, Flatten, Dropout, Activation
from keras.layers import Conv1D, MaxPooling1D, AveragePooling1D
from keras.models import Model
from keras.callbacks import ModelCheckpoint
from tensorflow.keras.optimizers import SGD
from fastapi import FastAPI, Request, UploadFile, File


import warnings
if not sys.warnoptions:
    warnings.simplefilter("ignore")
warnings.filterwarnings("ignore", category=DeprecationWarning)
os.environ['SENTENCE_TRANSFORMERS_HOME'] = './.cache'
model=from_pretrained_keras( 'Mohamed41/MODEL_EMOTION_AR_TEXT_72P')


def feat_ext(data):
    # Time_domain_features
    # ZCR Persody features or Low level ascoustic features
    result = np.array([])
    zcr = np.mean(librosa.feature.zero_crossing_rate(y=data).T, axis=0)
    result = np.hstack((result, zcr))  # stacking horizontally
    # Frequency_domain_features
    # Spectral and wavelet Features
    # MFCC
    mfcc = np.mean(librosa.feature.mfcc(y=data, sr=22050, n_mfcc=40).T, axis=0)
    result = np.hstack((result, mfcc))  # stacking horizontally
    return result


scaler = joblib.load('scaler.joblib')
encoder = joblib.load('encoder.joblib')


def get_predict_feat(path):
    d, s_rate = librosa.load(path, duration=2.5, offset=0.6)
    res = feat_ext(d)
    result = np.array(res)
    result = np.reshape(result, newshape=(1, 41))
    i_result = scaler.transform(result)
    final_result = np.expand_dims(i_result, axis=2)

    return final_result


emotions1 = {1: 'Neutral', 2: 'Calm', 3: 'Happy', 4: 'Sad',
             5: 'Angry', 6: 'Fear', 7: 'Disgust', 8: 'Surprise'}


def prediction(path1):
    res = get_predict_feat(path1)
    predictions = model.predict(res)
    y_pred = encoder.inverse_transform(predictions)
    return y_pred[0][0]


app = FastAPI()


@app.post("/")
async def read_root(request: Request, file: UploadFile = File(...)):
    json_data = await request.json()

    return {"filename": file.filename, "filepath": f"/app/{file.filename}"}