|
import type { TextGenerationStreamOutput } from "@huggingface/inference"; |
|
import type OpenAI from "openai"; |
|
import type { Stream } from "openai/streaming"; |
|
|
|
|
|
|
|
|
|
export async function* openAICompletionToTextGenerationStream( |
|
completionStream: Stream<OpenAI.Completions.Completion> |
|
) { |
|
let generatedText = ""; |
|
let tokenId = 0; |
|
for await (const completion of completionStream) { |
|
const { choices } = completion; |
|
const text = choices[0]?.text ?? ""; |
|
const last = choices[0]?.finish_reason === "stop" || choices[0]?.finish_reason === "length"; |
|
if (text) { |
|
generatedText = generatedText + text; |
|
} |
|
const output: TextGenerationStreamOutput = { |
|
token: { |
|
id: tokenId++, |
|
text, |
|
logprob: 0, |
|
special: last, |
|
}, |
|
generated_text: last ? generatedText : null, |
|
details: null, |
|
}; |
|
yield output; |
|
} |
|
} |
|
|