534 lines
		
	
	
		
			27 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
	
	
			
		
		
	
	
			534 lines
		
	
	
		
			27 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
	
	
| "use strict";
 | |
| // File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
 | |
| var _OpenAI_instances, _a, _OpenAI_encoder, _OpenAI_baseURLOverridden;
 | |
| Object.defineProperty(exports, "__esModule", { value: true });
 | |
| exports.OpenAI = void 0;
 | |
| const tslib_1 = require("./internal/tslib.js");
 | |
| const uuid_1 = require("./internal/utils/uuid.js");
 | |
| const values_1 = require("./internal/utils/values.js");
 | |
| const sleep_1 = require("./internal/utils/sleep.js");
 | |
| const errors_1 = require("./internal/errors.js");
 | |
| const detect_platform_1 = require("./internal/detect-platform.js");
 | |
| const Shims = tslib_1.__importStar(require("./internal/shims.js"));
 | |
| const Opts = tslib_1.__importStar(require("./internal/request-options.js"));
 | |
| const qs = tslib_1.__importStar(require("./internal/qs/index.js"));
 | |
| const version_1 = require("./version.js");
 | |
| const Errors = tslib_1.__importStar(require("./core/error.js"));
 | |
| const Pagination = tslib_1.__importStar(require("./core/pagination.js"));
 | |
| const Uploads = tslib_1.__importStar(require("./core/uploads.js"));
 | |
| const API = tslib_1.__importStar(require("./resources/index.js"));
 | |
| const api_promise_1 = require("./core/api-promise.js");
 | |
| const batches_1 = require("./resources/batches.js");
 | |
| const completions_1 = require("./resources/completions.js");
 | |
| const embeddings_1 = require("./resources/embeddings.js");
 | |
| const files_1 = require("./resources/files.js");
 | |
| const images_1 = require("./resources/images.js");
 | |
| const models_1 = require("./resources/models.js");
 | |
| const moderations_1 = require("./resources/moderations.js");
 | |
| const webhooks_1 = require("./resources/webhooks.js");
 | |
| const audio_1 = require("./resources/audio/audio.js");
 | |
| const beta_1 = require("./resources/beta/beta.js");
 | |
| const chat_1 = require("./resources/chat/chat.js");
 | |
| const containers_1 = require("./resources/containers/containers.js");
 | |
| const conversations_1 = require("./resources/conversations/conversations.js");
 | |
| const evals_1 = require("./resources/evals/evals.js");
 | |
| const fine_tuning_1 = require("./resources/fine-tuning/fine-tuning.js");
 | |
| const graders_1 = require("./resources/graders/graders.js");
 | |
| const responses_1 = require("./resources/responses/responses.js");
 | |
| const uploads_1 = require("./resources/uploads/uploads.js");
 | |
| const vector_stores_1 = require("./resources/vector-stores/vector-stores.js");
 | |
| const detect_platform_2 = require("./internal/detect-platform.js");
 | |
| const headers_1 = require("./internal/headers.js");
 | |
| const env_1 = require("./internal/utils/env.js");
 | |
| const log_1 = require("./internal/utils/log.js");
 | |
| const values_2 = require("./internal/utils/values.js");
 | |
| /**
 | |
|  * API Client for interfacing with the OpenAI API.
 | |
|  */
 | |
| class OpenAI {
 | |
|     /**
 | |
|      * API Client for interfacing with the OpenAI API.
 | |
|      *
 | |
|      * @param {string | undefined} [opts.apiKey=process.env['OPENAI_API_KEY'] ?? undefined]
 | |
|      * @param {string | null | undefined} [opts.organization=process.env['OPENAI_ORG_ID'] ?? null]
 | |
|      * @param {string | null | undefined} [opts.project=process.env['OPENAI_PROJECT_ID'] ?? null]
 | |
|      * @param {string | null | undefined} [opts.webhookSecret=process.env['OPENAI_WEBHOOK_SECRET'] ?? null]
 | |
|      * @param {string} [opts.baseURL=process.env['OPENAI_BASE_URL'] ?? https://api.openai.com/v1] - Override the default base URL for the API.
 | |
|      * @param {number} [opts.timeout=10 minutes] - The maximum amount of time (in milliseconds) the client will wait for a response before timing out.
 | |
|      * @param {MergedRequestInit} [opts.fetchOptions] - Additional `RequestInit` options to be passed to `fetch` calls.
 | |
|      * @param {Fetch} [opts.fetch] - Specify a custom `fetch` function implementation.
 | |
|      * @param {number} [opts.maxRetries=2] - The maximum number of times the client will retry a request.
 | |
|      * @param {HeadersLike} opts.defaultHeaders - Default headers to include with every request to the API.
 | |
|      * @param {Record<string, string | undefined>} opts.defaultQuery - Default query parameters to include with every request to the API.
 | |
|      * @param {boolean} [opts.dangerouslyAllowBrowser=false] - By default, client-side use of this library is not allowed, as it risks exposing your secret API credentials to attackers.
 | |
|      */
 | |
|     constructor({ baseURL = (0, env_1.readEnv)('OPENAI_BASE_URL'), apiKey = (0, env_1.readEnv)('OPENAI_API_KEY'), organization = (0, env_1.readEnv)('OPENAI_ORG_ID') ?? null, project = (0, env_1.readEnv)('OPENAI_PROJECT_ID') ?? null, webhookSecret = (0, env_1.readEnv)('OPENAI_WEBHOOK_SECRET') ?? null, ...opts } = {}) {
 | |
|         _OpenAI_instances.add(this);
 | |
|         _OpenAI_encoder.set(this, void 0);
 | |
|         this.completions = new API.Completions(this);
 | |
|         this.chat = new API.Chat(this);
 | |
|         this.embeddings = new API.Embeddings(this);
 | |
|         this.files = new API.Files(this);
 | |
|         this.images = new API.Images(this);
 | |
|         this.audio = new API.Audio(this);
 | |
|         this.moderations = new API.Moderations(this);
 | |
|         this.models = new API.Models(this);
 | |
|         this.fineTuning = new API.FineTuning(this);
 | |
|         this.graders = new API.Graders(this);
 | |
|         this.vectorStores = new API.VectorStores(this);
 | |
|         this.webhooks = new API.Webhooks(this);
 | |
|         this.beta = new API.Beta(this);
 | |
|         this.batches = new API.Batches(this);
 | |
|         this.uploads = new API.Uploads(this);
 | |
|         this.responses = new API.Responses(this);
 | |
|         this.conversations = new API.Conversations(this);
 | |
|         this.evals = new API.Evals(this);
 | |
|         this.containers = new API.Containers(this);
 | |
|         if (apiKey === undefined) {
 | |
|             throw new Errors.OpenAIError("The OPENAI_API_KEY environment variable is missing or empty; either provide it, or instantiate the OpenAI client with an apiKey option, like new OpenAI({ apiKey: 'My API Key' }).");
 | |
|         }
 | |
|         const options = {
 | |
|             apiKey,
 | |
|             organization,
 | |
|             project,
 | |
|             webhookSecret,
 | |
|             ...opts,
 | |
|             baseURL: baseURL || `https://api.openai.com/v1`,
 | |
|         };
 | |
|         if (!options.dangerouslyAllowBrowser && (0, detect_platform_2.isRunningInBrowser)()) {
 | |
|             throw new Errors.OpenAIError("It looks like you're running in a browser-like environment.\n\nThis is disabled by default, as it risks exposing your secret API credentials to attackers.\nIf you understand the risks and have appropriate mitigations in place,\nyou can set the `dangerouslyAllowBrowser` option to `true`, e.g.,\n\nnew OpenAI({ apiKey, dangerouslyAllowBrowser: true });\n\nhttps://help.openai.com/en/articles/5112595-best-practices-for-api-key-safety\n");
 | |
|         }
 | |
|         this.baseURL = options.baseURL;
 | |
|         this.timeout = options.timeout ?? _a.DEFAULT_TIMEOUT /* 10 minutes */;
 | |
|         this.logger = options.logger ?? console;
 | |
|         const defaultLogLevel = 'warn';
 | |
|         // Set default logLevel early so that we can log a warning in parseLogLevel.
 | |
|         this.logLevel = defaultLogLevel;
 | |
|         this.logLevel =
 | |
|             (0, log_1.parseLogLevel)(options.logLevel, 'ClientOptions.logLevel', this) ??
 | |
|                 (0, log_1.parseLogLevel)((0, env_1.readEnv)('OPENAI_LOG'), "process.env['OPENAI_LOG']", this) ??
 | |
|                 defaultLogLevel;
 | |
|         this.fetchOptions = options.fetchOptions;
 | |
|         this.maxRetries = options.maxRetries ?? 2;
 | |
|         this.fetch = options.fetch ?? Shims.getDefaultFetch();
 | |
|         tslib_1.__classPrivateFieldSet(this, _OpenAI_encoder, Opts.FallbackEncoder, "f");
 | |
|         this._options = options;
 | |
|         this.apiKey = apiKey;
 | |
|         this.organization = organization;
 | |
|         this.project = project;
 | |
|         this.webhookSecret = webhookSecret;
 | |
|     }
 | |
|     /**
 | |
|      * Create a new client instance re-using the same options given to the current client with optional overriding.
 | |
|      */
 | |
|     withOptions(options) {
 | |
|         const client = new this.constructor({
 | |
|             ...this._options,
 | |
|             baseURL: this.baseURL,
 | |
|             maxRetries: this.maxRetries,
 | |
|             timeout: this.timeout,
 | |
|             logger: this.logger,
 | |
|             logLevel: this.logLevel,
 | |
|             fetch: this.fetch,
 | |
|             fetchOptions: this.fetchOptions,
 | |
|             apiKey: this.apiKey,
 | |
|             organization: this.organization,
 | |
|             project: this.project,
 | |
|             webhookSecret: this.webhookSecret,
 | |
|             ...options,
 | |
|         });
 | |
|         return client;
 | |
|     }
 | |
|     defaultQuery() {
 | |
|         return this._options.defaultQuery;
 | |
|     }
 | |
|     validateHeaders({ values, nulls }) {
 | |
|         return;
 | |
|     }
 | |
|     async authHeaders(opts) {
 | |
|         return (0, headers_1.buildHeaders)([{ Authorization: `Bearer ${this.apiKey}` }]);
 | |
|     }
 | |
|     stringifyQuery(query) {
 | |
|         return qs.stringify(query, { arrayFormat: 'brackets' });
 | |
|     }
 | |
|     getUserAgent() {
 | |
|         return `${this.constructor.name}/JS ${version_1.VERSION}`;
 | |
|     }
 | |
|     defaultIdempotencyKey() {
 | |
|         return `stainless-node-retry-${(0, uuid_1.uuid4)()}`;
 | |
|     }
 | |
|     makeStatusError(status, error, message, headers) {
 | |
|         return Errors.APIError.generate(status, error, message, headers);
 | |
|     }
 | |
|     buildURL(path, query, defaultBaseURL) {
 | |
|         const baseURL = (!tslib_1.__classPrivateFieldGet(this, _OpenAI_instances, "m", _OpenAI_baseURLOverridden).call(this) && defaultBaseURL) || this.baseURL;
 | |
|         const url = (0, values_1.isAbsoluteURL)(path) ?
 | |
|             new URL(path)
 | |
|             : new URL(baseURL + (baseURL.endsWith('/') && path.startsWith('/') ? path.slice(1) : path));
 | |
|         const defaultQuery = this.defaultQuery();
 | |
|         if (!(0, values_2.isEmptyObj)(defaultQuery)) {
 | |
|             query = { ...defaultQuery, ...query };
 | |
|         }
 | |
|         if (typeof query === 'object' && query && !Array.isArray(query)) {
 | |
|             url.search = this.stringifyQuery(query);
 | |
|         }
 | |
|         return url.toString();
 | |
|     }
 | |
|     /**
 | |
|      * Used as a callback for mutating the given `FinalRequestOptions` object.
 | |
|      */
 | |
|     async prepareOptions(options) { }
 | |
|     /**
 | |
|      * Used as a callback for mutating the given `RequestInit` object.
 | |
|      *
 | |
|      * This is useful for cases where you want to add certain headers based off of
 | |
|      * the request properties, e.g. `method` or `url`.
 | |
|      */
 | |
|     async prepareRequest(request, { url, options }) { }
 | |
|     get(path, opts) {
 | |
|         return this.methodRequest('get', path, opts);
 | |
|     }
 | |
|     post(path, opts) {
 | |
|         return this.methodRequest('post', path, opts);
 | |
|     }
 | |
|     patch(path, opts) {
 | |
|         return this.methodRequest('patch', path, opts);
 | |
|     }
 | |
|     put(path, opts) {
 | |
|         return this.methodRequest('put', path, opts);
 | |
|     }
 | |
|     delete(path, opts) {
 | |
|         return this.methodRequest('delete', path, opts);
 | |
|     }
 | |
|     methodRequest(method, path, opts) {
 | |
|         return this.request(Promise.resolve(opts).then((opts) => {
 | |
|             return { method, path, ...opts };
 | |
|         }));
 | |
|     }
 | |
|     request(options, remainingRetries = null) {
 | |
|         return new api_promise_1.APIPromise(this, this.makeRequest(options, remainingRetries, undefined));
 | |
|     }
 | |
|     async makeRequest(optionsInput, retriesRemaining, retryOfRequestLogID) {
 | |
|         const options = await optionsInput;
 | |
|         const maxRetries = options.maxRetries ?? this.maxRetries;
 | |
|         if (retriesRemaining == null) {
 | |
|             retriesRemaining = maxRetries;
 | |
|         }
 | |
|         await this.prepareOptions(options);
 | |
|         const { req, url, timeout } = await this.buildRequest(options, {
 | |
|             retryCount: maxRetries - retriesRemaining,
 | |
|         });
 | |
|         await this.prepareRequest(req, { url, options });
 | |
|         /** Not an API request ID, just for correlating local log entries. */
 | |
|         const requestLogID = 'log_' + ((Math.random() * (1 << 24)) | 0).toString(16).padStart(6, '0');
 | |
|         const retryLogStr = retryOfRequestLogID === undefined ? '' : `, retryOf: ${retryOfRequestLogID}`;
 | |
|         const startTime = Date.now();
 | |
|         (0, log_1.loggerFor)(this).debug(`[${requestLogID}] sending request`, (0, log_1.formatRequestDetails)({
 | |
|             retryOfRequestLogID,
 | |
|             method: options.method,
 | |
|             url,
 | |
|             options,
 | |
|             headers: req.headers,
 | |
|         }));
 | |
|         if (options.signal?.aborted) {
 | |
|             throw new Errors.APIUserAbortError();
 | |
|         }
 | |
|         const controller = new AbortController();
 | |
|         const response = await this.fetchWithTimeout(url, req, timeout, controller).catch(errors_1.castToError);
 | |
|         const headersTime = Date.now();
 | |
|         if (response instanceof Error) {
 | |
|             const retryMessage = `retrying, ${retriesRemaining} attempts remaining`;
 | |
|             if (options.signal?.aborted) {
 | |
|                 throw new Errors.APIUserAbortError();
 | |
|             }
 | |
|             // detect native connection timeout errors
 | |
|             // deno throws "TypeError: error sending request for url (https://example/): client error (Connect): tcp connect error: Operation timed out (os error 60): Operation timed out (os error 60)"
 | |
|             // undici throws "TypeError: fetch failed" with cause "ConnectTimeoutError: Connect Timeout Error (attempted address: example:443, timeout: 1ms)"
 | |
|             // others do not provide enough information to distinguish timeouts from other connection errors
 | |
|             const isTimeout = (0, errors_1.isAbortError)(response) ||
 | |
|                 /timed? ?out/i.test(String(response) + ('cause' in response ? String(response.cause) : ''));
 | |
|             if (retriesRemaining) {
 | |
|                 (0, log_1.loggerFor)(this).info(`[${requestLogID}] connection ${isTimeout ? 'timed out' : 'failed'} - ${retryMessage}`);
 | |
|                 (0, log_1.loggerFor)(this).debug(`[${requestLogID}] connection ${isTimeout ? 'timed out' : 'failed'} (${retryMessage})`, (0, log_1.formatRequestDetails)({
 | |
|                     retryOfRequestLogID,
 | |
|                     url,
 | |
|                     durationMs: headersTime - startTime,
 | |
|                     message: response.message,
 | |
|                 }));
 | |
|                 return this.retryRequest(options, retriesRemaining, retryOfRequestLogID ?? requestLogID);
 | |
|             }
 | |
|             (0, log_1.loggerFor)(this).info(`[${requestLogID}] connection ${isTimeout ? 'timed out' : 'failed'} - error; no more retries left`);
 | |
|             (0, log_1.loggerFor)(this).debug(`[${requestLogID}] connection ${isTimeout ? 'timed out' : 'failed'} (error; no more retries left)`, (0, log_1.formatRequestDetails)({
 | |
|                 retryOfRequestLogID,
 | |
|                 url,
 | |
|                 durationMs: headersTime - startTime,
 | |
|                 message: response.message,
 | |
|             }));
 | |
|             if (isTimeout) {
 | |
|                 throw new Errors.APIConnectionTimeoutError();
 | |
|             }
 | |
|             throw new Errors.APIConnectionError({ cause: response });
 | |
|         }
 | |
|         const specialHeaders = [...response.headers.entries()]
 | |
|             .filter(([name]) => name === 'x-request-id')
 | |
|             .map(([name, value]) => ', ' + name + ': ' + JSON.stringify(value))
 | |
|             .join('');
 | |
|         const responseInfo = `[${requestLogID}${retryLogStr}${specialHeaders}] ${req.method} ${url} ${response.ok ? 'succeeded' : 'failed'} with status ${response.status} in ${headersTime - startTime}ms`;
 | |
|         if (!response.ok) {
 | |
|             const shouldRetry = await this.shouldRetry(response);
 | |
|             if (retriesRemaining && shouldRetry) {
 | |
|                 const retryMessage = `retrying, ${retriesRemaining} attempts remaining`;
 | |
|                 // We don't need the body of this response.
 | |
|                 await Shims.CancelReadableStream(response.body);
 | |
|                 (0, log_1.loggerFor)(this).info(`${responseInfo} - ${retryMessage}`);
 | |
|                 (0, log_1.loggerFor)(this).debug(`[${requestLogID}] response error (${retryMessage})`, (0, log_1.formatRequestDetails)({
 | |
|                     retryOfRequestLogID,
 | |
|                     url: response.url,
 | |
|                     status: response.status,
 | |
|                     headers: response.headers,
 | |
|                     durationMs: headersTime - startTime,
 | |
|                 }));
 | |
|                 return this.retryRequest(options, retriesRemaining, retryOfRequestLogID ?? requestLogID, response.headers);
 | |
|             }
 | |
|             const retryMessage = shouldRetry ? `error; no more retries left` : `error; not retryable`;
 | |
|             (0, log_1.loggerFor)(this).info(`${responseInfo} - ${retryMessage}`);
 | |
|             const errText = await response.text().catch((err) => (0, errors_1.castToError)(err).message);
 | |
|             const errJSON = (0, values_1.safeJSON)(errText);
 | |
|             const errMessage = errJSON ? undefined : errText;
 | |
|             (0, log_1.loggerFor)(this).debug(`[${requestLogID}] response error (${retryMessage})`, (0, log_1.formatRequestDetails)({
 | |
|                 retryOfRequestLogID,
 | |
|                 url: response.url,
 | |
|                 status: response.status,
 | |
|                 headers: response.headers,
 | |
|                 message: errMessage,
 | |
|                 durationMs: Date.now() - startTime,
 | |
|             }));
 | |
|             const err = this.makeStatusError(response.status, errJSON, errMessage, response.headers);
 | |
|             throw err;
 | |
|         }
 | |
|         (0, log_1.loggerFor)(this).info(responseInfo);
 | |
|         (0, log_1.loggerFor)(this).debug(`[${requestLogID}] response start`, (0, log_1.formatRequestDetails)({
 | |
|             retryOfRequestLogID,
 | |
|             url: response.url,
 | |
|             status: response.status,
 | |
|             headers: response.headers,
 | |
|             durationMs: headersTime - startTime,
 | |
|         }));
 | |
|         return { response, options, controller, requestLogID, retryOfRequestLogID, startTime };
 | |
|     }
 | |
|     getAPIList(path, Page, opts) {
 | |
|         return this.requestAPIList(Page, { method: 'get', path, ...opts });
 | |
|     }
 | |
|     requestAPIList(Page, options) {
 | |
|         const request = this.makeRequest(options, null, undefined);
 | |
|         return new Pagination.PagePromise(this, request, Page);
 | |
|     }
 | |
|     async fetchWithTimeout(url, init, ms, controller) {
 | |
|         const { signal, method, ...options } = init || {};
 | |
|         if (signal)
 | |
|             signal.addEventListener('abort', () => controller.abort());
 | |
|         const timeout = setTimeout(() => controller.abort(), ms);
 | |
|         const isReadableBody = (globalThis.ReadableStream && options.body instanceof globalThis.ReadableStream) ||
 | |
|             (typeof options.body === 'object' && options.body !== null && Symbol.asyncIterator in options.body);
 | |
|         const fetchOptions = {
 | |
|             signal: controller.signal,
 | |
|             ...(isReadableBody ? { duplex: 'half' } : {}),
 | |
|             method: 'GET',
 | |
|             ...options,
 | |
|         };
 | |
|         if (method) {
 | |
|             // Custom methods like 'patch' need to be uppercased
 | |
|             // See https://github.com/nodejs/undici/issues/2294
 | |
|             fetchOptions.method = method.toUpperCase();
 | |
|         }
 | |
|         try {
 | |
|             // use undefined this binding; fetch errors if bound to something else in browser/cloudflare
 | |
|             return await this.fetch.call(undefined, url, fetchOptions);
 | |
|         }
 | |
|         finally {
 | |
|             clearTimeout(timeout);
 | |
|         }
 | |
|     }
 | |
|     async shouldRetry(response) {
 | |
|         // Note this is not a standard header.
 | |
|         const shouldRetryHeader = response.headers.get('x-should-retry');
 | |
|         // If the server explicitly says whether or not to retry, obey.
 | |
|         if (shouldRetryHeader === 'true')
 | |
|             return true;
 | |
|         if (shouldRetryHeader === 'false')
 | |
|             return false;
 | |
|         // Retry on request timeouts.
 | |
|         if (response.status === 408)
 | |
|             return true;
 | |
|         // Retry on lock timeouts.
 | |
|         if (response.status === 409)
 | |
|             return true;
 | |
|         // Retry on rate limits.
 | |
|         if (response.status === 429)
 | |
|             return true;
 | |
|         // Retry internal errors.
 | |
|         if (response.status >= 500)
 | |
|             return true;
 | |
|         return false;
 | |
|     }
 | |
|     async retryRequest(options, retriesRemaining, requestLogID, responseHeaders) {
 | |
|         let timeoutMillis;
 | |
|         // Note the `retry-after-ms` header may not be standard, but is a good idea and we'd like proactive support for it.
 | |
|         const retryAfterMillisHeader = responseHeaders?.get('retry-after-ms');
 | |
|         if (retryAfterMillisHeader) {
 | |
|             const timeoutMs = parseFloat(retryAfterMillisHeader);
 | |
|             if (!Number.isNaN(timeoutMs)) {
 | |
|                 timeoutMillis = timeoutMs;
 | |
|             }
 | |
|         }
 | |
|         // About the Retry-After header: https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/Retry-After
 | |
|         const retryAfterHeader = responseHeaders?.get('retry-after');
 | |
|         if (retryAfterHeader && !timeoutMillis) {
 | |
|             const timeoutSeconds = parseFloat(retryAfterHeader);
 | |
|             if (!Number.isNaN(timeoutSeconds)) {
 | |
|                 timeoutMillis = timeoutSeconds * 1000;
 | |
|             }
 | |
|             else {
 | |
|                 timeoutMillis = Date.parse(retryAfterHeader) - Date.now();
 | |
|             }
 | |
|         }
 | |
|         // If the API asks us to wait a certain amount of time (and it's a reasonable amount),
 | |
|         // just do what it says, but otherwise calculate a default
 | |
|         if (!(timeoutMillis && 0 <= timeoutMillis && timeoutMillis < 60 * 1000)) {
 | |
|             const maxRetries = options.maxRetries ?? this.maxRetries;
 | |
|             timeoutMillis = this.calculateDefaultRetryTimeoutMillis(retriesRemaining, maxRetries);
 | |
|         }
 | |
|         await (0, sleep_1.sleep)(timeoutMillis);
 | |
|         return this.makeRequest(options, retriesRemaining - 1, requestLogID);
 | |
|     }
 | |
|     calculateDefaultRetryTimeoutMillis(retriesRemaining, maxRetries) {
 | |
|         const initialRetryDelay = 0.5;
 | |
|         const maxRetryDelay = 8.0;
 | |
|         const numRetries = maxRetries - retriesRemaining;
 | |
|         // Apply exponential backoff, but not more than the max.
 | |
|         const sleepSeconds = Math.min(initialRetryDelay * Math.pow(2, numRetries), maxRetryDelay);
 | |
|         // Apply some jitter, take up to at most 25 percent of the retry time.
 | |
|         const jitter = 1 - Math.random() * 0.25;
 | |
|         return sleepSeconds * jitter * 1000;
 | |
|     }
 | |
|     async buildRequest(inputOptions, { retryCount = 0 } = {}) {
 | |
|         const options = { ...inputOptions };
 | |
|         const { method, path, query, defaultBaseURL } = options;
 | |
|         const url = this.buildURL(path, query, defaultBaseURL);
 | |
|         if ('timeout' in options)
 | |
|             (0, values_1.validatePositiveInteger)('timeout', options.timeout);
 | |
|         options.timeout = options.timeout ?? this.timeout;
 | |
|         const { bodyHeaders, body } = this.buildBody({ options });
 | |
|         const reqHeaders = await this.buildHeaders({ options: inputOptions, method, bodyHeaders, retryCount });
 | |
|         const req = {
 | |
|             method,
 | |
|             headers: reqHeaders,
 | |
|             ...(options.signal && { signal: options.signal }),
 | |
|             ...(globalThis.ReadableStream &&
 | |
|                 body instanceof globalThis.ReadableStream && { duplex: 'half' }),
 | |
|             ...(body && { body }),
 | |
|             ...(this.fetchOptions ?? {}),
 | |
|             ...(options.fetchOptions ?? {}),
 | |
|         };
 | |
|         return { req, url, timeout: options.timeout };
 | |
|     }
 | |
|     async buildHeaders({ options, method, bodyHeaders, retryCount, }) {
 | |
|         let idempotencyHeaders = {};
 | |
|         if (this.idempotencyHeader && method !== 'get') {
 | |
|             if (!options.idempotencyKey)
 | |
|                 options.idempotencyKey = this.defaultIdempotencyKey();
 | |
|             idempotencyHeaders[this.idempotencyHeader] = options.idempotencyKey;
 | |
|         }
 | |
|         const headers = (0, headers_1.buildHeaders)([
 | |
|             idempotencyHeaders,
 | |
|             {
 | |
|                 Accept: 'application/json',
 | |
|                 'User-Agent': this.getUserAgent(),
 | |
|                 'X-Stainless-Retry-Count': String(retryCount),
 | |
|                 ...(options.timeout ? { 'X-Stainless-Timeout': String(Math.trunc(options.timeout / 1000)) } : {}),
 | |
|                 ...(0, detect_platform_1.getPlatformHeaders)(),
 | |
|                 'OpenAI-Organization': this.organization,
 | |
|                 'OpenAI-Project': this.project,
 | |
|             },
 | |
|             await this.authHeaders(options),
 | |
|             this._options.defaultHeaders,
 | |
|             bodyHeaders,
 | |
|             options.headers,
 | |
|         ]);
 | |
|         this.validateHeaders(headers);
 | |
|         return headers.values;
 | |
|     }
 | |
|     buildBody({ options: { body, headers: rawHeaders } }) {
 | |
|         if (!body) {
 | |
|             return { bodyHeaders: undefined, body: undefined };
 | |
|         }
 | |
|         const headers = (0, headers_1.buildHeaders)([rawHeaders]);
 | |
|         if (
 | |
|         // Pass raw type verbatim
 | |
|         ArrayBuffer.isView(body) ||
 | |
|             body instanceof ArrayBuffer ||
 | |
|             body instanceof DataView ||
 | |
|             (typeof body === 'string' &&
 | |
|                 // Preserve legacy string encoding behavior for now
 | |
|                 headers.values.has('content-type')) ||
 | |
|             // `Blob` is superset of `File`
 | |
|             (globalThis.Blob && body instanceof globalThis.Blob) ||
 | |
|             // `FormData` -> `multipart/form-data`
 | |
|             body instanceof FormData ||
 | |
|             // `URLSearchParams` -> `application/x-www-form-urlencoded`
 | |
|             body instanceof URLSearchParams ||
 | |
|             // Send chunked stream (each chunk has own `length`)
 | |
|             (globalThis.ReadableStream && body instanceof globalThis.ReadableStream)) {
 | |
|             return { bodyHeaders: undefined, body: body };
 | |
|         }
 | |
|         else if (typeof body === 'object' &&
 | |
|             (Symbol.asyncIterator in body ||
 | |
|                 (Symbol.iterator in body && 'next' in body && typeof body.next === 'function'))) {
 | |
|             return { bodyHeaders: undefined, body: Shims.ReadableStreamFrom(body) };
 | |
|         }
 | |
|         else {
 | |
|             return tslib_1.__classPrivateFieldGet(this, _OpenAI_encoder, "f").call(this, { body, headers });
 | |
|         }
 | |
|     }
 | |
| }
 | |
| exports.OpenAI = OpenAI;
 | |
| _a = OpenAI, _OpenAI_encoder = new WeakMap(), _OpenAI_instances = new WeakSet(), _OpenAI_baseURLOverridden = function _OpenAI_baseURLOverridden() {
 | |
|     return this.baseURL !== 'https://api.openai.com/v1';
 | |
| };
 | |
| OpenAI.OpenAI = _a;
 | |
| OpenAI.DEFAULT_TIMEOUT = 600000; // 10 minutes
 | |
| OpenAI.OpenAIError = Errors.OpenAIError;
 | |
| OpenAI.APIError = Errors.APIError;
 | |
| OpenAI.APIConnectionError = Errors.APIConnectionError;
 | |
| OpenAI.APIConnectionTimeoutError = Errors.APIConnectionTimeoutError;
 | |
| OpenAI.APIUserAbortError = Errors.APIUserAbortError;
 | |
| OpenAI.NotFoundError = Errors.NotFoundError;
 | |
| OpenAI.ConflictError = Errors.ConflictError;
 | |
| OpenAI.RateLimitError = Errors.RateLimitError;
 | |
| OpenAI.BadRequestError = Errors.BadRequestError;
 | |
| OpenAI.AuthenticationError = Errors.AuthenticationError;
 | |
| OpenAI.InternalServerError = Errors.InternalServerError;
 | |
| OpenAI.PermissionDeniedError = Errors.PermissionDeniedError;
 | |
| OpenAI.UnprocessableEntityError = Errors.UnprocessableEntityError;
 | |
| OpenAI.InvalidWebhookSignatureError = Errors.InvalidWebhookSignatureError;
 | |
| OpenAI.toFile = Uploads.toFile;
 | |
| OpenAI.Completions = completions_1.Completions;
 | |
| OpenAI.Chat = chat_1.Chat;
 | |
| OpenAI.Embeddings = embeddings_1.Embeddings;
 | |
| OpenAI.Files = files_1.Files;
 | |
| OpenAI.Images = images_1.Images;
 | |
| OpenAI.Audio = audio_1.Audio;
 | |
| OpenAI.Moderations = moderations_1.Moderations;
 | |
| OpenAI.Models = models_1.Models;
 | |
| OpenAI.FineTuning = fine_tuning_1.FineTuning;
 | |
| OpenAI.Graders = graders_1.Graders;
 | |
| OpenAI.VectorStores = vector_stores_1.VectorStores;
 | |
| OpenAI.Webhooks = webhooks_1.Webhooks;
 | |
| OpenAI.Beta = beta_1.Beta;
 | |
| OpenAI.Batches = batches_1.Batches;
 | |
| OpenAI.Uploads = uploads_1.Uploads;
 | |
| OpenAI.Responses = responses_1.Responses;
 | |
| OpenAI.Conversations = conversations_1.Conversations;
 | |
| OpenAI.Evals = evals_1.Evals;
 | |
| OpenAI.Containers = containers_1.Containers;
 | |
| //# sourceMappingURL=client.js.map
 |