Thamaraikannan commited on
Commit
60c771d
·
verified ·
1 Parent(s): b355e83

Update helpers/tts.js

Browse files
Files changed (1) hide show
  1. helpers/tts.js +48 -65
helpers/tts.js CHANGED
@@ -33,70 +33,53 @@ if (!key || !region) {
33
  * @param {*} text text to convert to audio/speech
34
  * @param {*} filename optional - best for long text - temp file for converted speech/audio
35
  */
36
- const textToSpeech = async (text, voice)=> {
37
-
38
- // convert callback function to promise
39
- return new Promise((resolve, reject) => {
40
-
41
-
42
- let ssml = SSML.replace("__TEXT__", text);
43
-
44
-
45
- const speechConfig = sdk.SpeechConfig.fromSubscription(key, region);
46
- speechConfig.speechSynthesisOutputFormat = 5; // mp3
47
-
48
- let audioConfig = null;
49
-
50
- // if (filename) {
51
- let randomString = Math.random().toString(36).slice(2, 7);
52
- let filename = `./public/speech-${randomString}.mp3`;
53
- audioConfig = sdk.AudioConfig.fromAudioFileOutput(filename);
54
- // }
55
-
56
- let blendData = [];
57
- let timeStep = 1/60;
58
- let timeStamp = 0;
59
-
60
- const synthesizer = new sdk.SpeechSynthesizer(speechConfig, audioConfig);
61
-
62
- // Subscribes to viseme received event
63
- synthesizer.visemeReceived = function (s, e) {
64
-
65
- // `Animation` is an xml string for SVG or a json string for blend shapes
66
- var animation = JSON.parse(e.animation);
67
-
68
- _.each(animation.BlendShapes, blendArray => {
69
-
70
- let blend = {};
71
- _.each(blendShapeNames, (shapeName, i) => {
72
- blend[shapeName] = blendArray[i];
73
- });
74
-
75
- blendData.push({
76
- time: timeStamp,
77
- blendshapes: blend
78
- });
79
- console.log(`Timestamp: ${timeStamp.toFixed(3)}s`);
80
- console.log(JSON.stringify(blend, null, 2));
81
- timeStamp += timeStep;
82
- });
83
-
84
- }
85
-
86
-
87
- synthesizer.speakSsmlAsync(
88
- ssml,
89
- result => {
90
-
91
- synthesizer.close();
92
- resolve({blendData, filename: `/speech-${randomString}.mp3`});
93
-
94
- },
95
- error => {
96
- synthesizer.close();
97
- reject(error);
98
- });
99
- });
100
  };
101
 
102
- module.exports = textToSpeech;
 
33
  * @param {*} text text to convert to audio/speech
34
  * @param {*} filename optional - best for long text - temp file for converted speech/audio
35
  */
36
+ const textToSpeech = async (text, voice) => {
37
+ return new Promise((resolve, reject) => {
38
+ let ssml = SSML.replace("__TEXT__", text);
39
+
40
+ const speechConfig = sdk.SpeechConfig.fromSubscription(key, region);
41
+ speechConfig.speechSynthesisOutputFormat = 5; // mp3
42
+
43
+ let audioConfig = null;
44
+
45
+ let randomString = Math.random().toString(36).slice(2, 7);
46
+ let filename = `/tmp/speech-${randomString}.mp3`; // Write to /tmp directory
47
+ audioConfig = sdk.AudioConfig.fromAudioFileOutput(filename);
48
+
49
+ let blendData = [];
50
+ let timeStep = 1/60;
51
+ let timeStamp = 0;
52
+
53
+ const synthesizer = new sdk.SpeechSynthesizer(speechConfig, audioConfig);
54
+
55
+ synthesizer.visemeReceived = function (s, e) {
56
+ var animation = JSON.parse(e.animation);
57
+
58
+ _.each(animation.BlendShapes, blendArray => {
59
+ let blend = {};
60
+ _.each(blendShapeNames, (shapeName, i) => {
61
+ blend[shapeName] = blendArray[i];
62
+ });
63
+
64
+ blendData.push({
65
+ time: timeStamp,
66
+ blendshapes: blend
67
+ });
68
+ timeStamp += timeStep;
69
+ });
70
+ };
71
+
72
+ synthesizer.speakSsmlAsync(
73
+ ssml,
74
+ result => {
75
+ synthesizer.close();
76
+ resolve({ blendData, filename: `/speech-${randomString}.mp3` });
77
+ },
78
+ error => {
79
+ synthesizer.close();
80
+ reject(error);
81
+ }
82
+ );
83
+ });
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
84
  };
85