Spaces:
Build error
Build error
import * as Core from '@anthropic-ai/sdk/core'; | |
import { APIPromise } from '@anthropic-ai/sdk/core'; | |
import { APIResource } from '@anthropic-ai/sdk/resource'; | |
import * as CompletionsAPI from '@anthropic-ai/sdk/resources/completions'; | |
import { Stream } from '@anthropic-ai/sdk/streaming'; | |
export declare class Completions extends APIResource { | |
/** | |
* Create a completion | |
*/ | |
create(body: CompletionCreateParamsNonStreaming, options?: Core.RequestOptions): APIPromise<Completion>; | |
create(body: CompletionCreateParamsStreaming, options?: Core.RequestOptions): APIPromise<Stream<Completion>>; | |
create(body: CompletionCreateParamsBase, options?: Core.RequestOptions): APIPromise<Stream<Completion> | Completion>; | |
} | |
export interface Completion { | |
/** | |
* The resulting completion up to and excluding the stop sequences. | |
*/ | |
completion: string; | |
/** | |
* The model that performed the completion. | |
*/ | |
model: string; | |
/** | |
* The reason that we stopped sampling. | |
* | |
* This may be one the following values: | |
* | |
* - `"stop_sequence"`: we reached a stop sequence — either provided by you via the | |
* `stop_sequences` parameter, or a stop sequence built into the model | |
* - `"max_tokens"`: we exceeded `max_tokens_to_sample` or the model's maximum | |
*/ | |
stop_reason: string; | |
} | |
export type CompletionCreateParams = CompletionCreateParamsNonStreaming | CompletionCreateParamsStreaming; | |
export interface CompletionCreateParamsBase { | |
/** | |
* The maximum number of tokens to generate before stopping. | |
* | |
* Note that our models may stop _before_ reaching this maximum. This parameter | |
* only specifies the absolute maximum number of tokens to generate. | |
*/ | |
max_tokens_to_sample: number; | |
/** | |
* The model that will complete your prompt. | |
* | |
* As we improve Claude, we develop new versions of it that you can query. This | |
* parameter controls which version of Claude answers your request. Right now we | |
* are offering two model families: Claude, and Claude Instant. You can use them by | |
* setting `model` to `"claude-2"` or `"claude-instant-1"`, respectively. See | |
* [models](https://docs.anthropic.com/claude/reference/selecting-a-model) for | |
* additional details. | |
*/ | |
model: (string & {}) | 'claude-2' | 'claude-instant-1'; | |
/** | |
* The prompt that you want Claude to complete. | |
* | |
* For proper response generation you will need to format your prompt as follows: | |
* | |
* ```javascript | |
* const userQuestion = r"Why is the sky blue?"; | |
* const prompt = `\n\nHuman: ${userQuestion}\n\nAssistant:`; | |
* ``` | |
* | |
* See our | |
* [comments on prompts](https://docs.anthropic.com/claude/docs/introduction-to-prompt-design) | |
* for more context. | |
*/ | |
prompt: string; | |
/** | |
* An object describing metadata about the request. | |
*/ | |
metadata?: CompletionCreateParams.Metadata; | |
/** | |
* Sequences that will cause the model to stop generating completion text. | |
* | |
* Our models stop on `"\n\nHuman:"`, and may include additional built-in stop | |
* sequences in the future. By providing the stop_sequences parameter, you may | |
* include additional strings that will cause the model to stop generating. | |
*/ | |
stop_sequences?: Array<string>; | |
/** | |
* Whether to incrementally stream the response using server-sent events. | |
* | |
* See | |
* [this guide to SSE events](https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events/Using_server-sent_events) | |
* for details. | |
*/ | |
stream?: boolean; | |
/** | |
* Amount of randomness injected into the response. | |
* | |
* Defaults to 1. Ranges from 0 to 1. Use temp closer to 0 for analytical / | |
* multiple choice, and closer to 1 for creative and generative tasks. | |
*/ | |
temperature?: number; | |
/** | |
* Only sample from the top K options for each subsequent token. | |
* | |
* Used to remove "long tail" low probability responses. | |
* [Learn more technical details here](https://towardsdatascience.com/how-to-sample-from-language-models-682bceb97277). | |
*/ | |
top_k?: number; | |
/** | |
* Use nucleus sampling. | |
* | |
* In nucleus sampling, we compute the cumulative distribution over all the options | |
* for each subsequent token in decreasing probability order and cut it off once it | |
* reaches a particular probability specified by `top_p`. You should either alter | |
* `temperature` or `top_p`, but not both. | |
*/ | |
top_p?: number; | |
} | |
export declare namespace CompletionCreateParams { | |
/** | |
* An object describing metadata about the request. | |
*/ | |
interface Metadata { | |
/** | |
* An external identifier for the user who is associated with the request. | |
* | |
* This should be a uuid, hash value, or other opaque identifier. Anthropic may use | |
* this id to help detect abuse. Do not include any identifying information such as | |
* name, email address, or phone number. | |
*/ | |
user_id?: string; | |
} | |
type CompletionCreateParamsNonStreaming = CompletionsAPI.CompletionCreateParamsNonStreaming; | |
type CompletionCreateParamsStreaming = CompletionsAPI.CompletionCreateParamsStreaming; | |
} | |
export interface CompletionCreateParamsNonStreaming extends CompletionCreateParamsBase { | |
/** | |
* Whether to incrementally stream the response using server-sent events. | |
* | |
* See | |
* [this guide to SSE events](https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events/Using_server-sent_events) | |
* for details. | |
*/ | |
stream?: false; | |
} | |
export interface CompletionCreateParamsStreaming extends CompletionCreateParamsBase { | |
/** | |
* Whether to incrementally stream the response using server-sent events. | |
* | |
* See | |
* [this guide to SSE events](https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events/Using_server-sent_events) | |
* for details. | |
*/ | |
stream: true; | |
} | |
export declare namespace Completions { | |
export import Completion = CompletionsAPI.Completion; | |
export import CompletionCreateParams = CompletionsAPI.CompletionCreateParams; | |
export import CompletionCreateParamsNonStreaming = CompletionsAPI.CompletionCreateParamsNonStreaming; | |
export import CompletionCreateParamsStreaming = CompletionsAPI.CompletionCreateParamsStreaming; | |
} | |
//# sourceMappingURL=completions.d.ts.map |