145 lines
5.5 KiB
TypeScript
145 lines
5.5 KiB
TypeScript
import { describe, expect, it } from "vitest";
|
|
import { getModel } from "../src/models.js";
|
|
import { complete } from "../src/stream.js";
|
|
import type { Api, Context, Model, StreamOptions } from "../src/types.js";
|
|
import { hasAzureOpenAICredentials, resolveAzureDeploymentName } from "./azure-utils.js";
|
|
import { resolveApiKey } from "./oauth.js";
|
|
|
|
type StreamOptionsWithExtras = StreamOptions & Record<string, unknown>;
|
|
|
|
const oauthTokens = await Promise.all([
|
|
resolveApiKey("github-copilot"),
|
|
resolveApiKey("google-gemini-cli"),
|
|
resolveApiKey("google-antigravity"),
|
|
resolveApiKey("openai-codex"),
|
|
]);
|
|
const [githubCopilotToken, geminiCliToken, antigravityToken, openaiCodexToken] = oauthTokens;
|
|
|
|
async function expectResponseId<TApi extends Api>(model: Model<TApi>, options: StreamOptionsWithExtras = {}) {
|
|
const context: Context = {
|
|
systemPrompt: "You are a helpful assistant. Be concise.",
|
|
messages: [{ role: "user", content: "Reply with exactly: response id test", timestamp: Date.now() }],
|
|
};
|
|
|
|
const response = await complete(model, context, options);
|
|
|
|
expect(response.stopReason, response.errorMessage).not.toBe("error");
|
|
expect(response.responseId).toBeTruthy();
|
|
expect(typeof response.responseId).toBe("string");
|
|
}
|
|
|
|
describe("responseId E2E Tests", () => {
|
|
describe.skipIf(!process.env.GEMINI_API_KEY)("Google Provider", () => {
|
|
const llm = getModel("google", "gemini-2.5-flash");
|
|
|
|
it("should expose responseId", { retry: 3, timeout: 30000 }, async () => {
|
|
await expectResponseId(llm);
|
|
});
|
|
});
|
|
|
|
describe("Google Vertex Provider", () => {
|
|
const vertexProject = process.env.GOOGLE_CLOUD_PROJECT || process.env.GCLOUD_PROJECT;
|
|
const vertexLocation = process.env.GOOGLE_CLOUD_LOCATION;
|
|
const vertexApiKey = process.env.GOOGLE_CLOUD_API_KEY;
|
|
const isVertexConfigured = Boolean(vertexProject && vertexLocation);
|
|
const vertexOptions = { project: vertexProject, location: vertexLocation } as const;
|
|
const llm = getModel("google-vertex", "gemini-3-flash-preview");
|
|
|
|
it.skipIf(!isVertexConfigured)("should expose responseId with ADC", { retry: 3, timeout: 30000 }, async () => {
|
|
await expectResponseId(llm, vertexOptions);
|
|
});
|
|
|
|
it.skipIf(!vertexApiKey)("should expose responseId with API key", { retry: 3, timeout: 30000 }, async () => {
|
|
await expectResponseId(llm, { apiKey: vertexApiKey! });
|
|
});
|
|
});
|
|
|
|
describe.skipIf(!process.env.OPENAI_API_KEY)("OpenAI Completions Provider", () => {
|
|
const { compat: _compat, ...baseModel } = getModel("openai", "gpt-4o-mini");
|
|
void _compat;
|
|
const llm: Model<"openai-completions"> = {
|
|
...baseModel,
|
|
api: "openai-completions",
|
|
};
|
|
|
|
it("should expose responseId", { retry: 3, timeout: 30000 }, async () => {
|
|
await expectResponseId(llm);
|
|
});
|
|
});
|
|
|
|
describe.skipIf(!process.env.OPENAI_API_KEY)("OpenAI Responses Provider", () => {
|
|
const llm = getModel("openai", "gpt-5-mini");
|
|
|
|
it("should expose responseId", { retry: 3, timeout: 30000 }, async () => {
|
|
await expectResponseId(llm);
|
|
});
|
|
});
|
|
|
|
describe.skipIf(!process.env.ANTHROPIC_API_KEY)("Anthropic Provider", () => {
|
|
const llm = getModel("anthropic", "claude-sonnet-4-5");
|
|
|
|
it("should expose responseId", { retry: 3, timeout: 30000 }, async () => {
|
|
await expectResponseId(llm);
|
|
});
|
|
});
|
|
|
|
describe.skipIf(!hasAzureOpenAICredentials())("Azure OpenAI Responses Provider", () => {
|
|
const llm = getModel("azure-openai-responses", "gpt-4o-mini");
|
|
const azureDeploymentName = resolveAzureDeploymentName(llm.id);
|
|
const azureOptions = azureDeploymentName ? { azureDeploymentName } : {};
|
|
|
|
it("should expose responseId", { retry: 3, timeout: 30000 }, async () => {
|
|
await expectResponseId(llm, azureOptions);
|
|
});
|
|
});
|
|
|
|
describe.skipIf(!process.env.MISTRAL_API_KEY)("Mistral Provider", () => {
|
|
const llm = getModel("mistral", "devstral-medium-latest");
|
|
|
|
it("should expose responseId", { retry: 3, timeout: 30000 }, async () => {
|
|
await expectResponseId(llm);
|
|
});
|
|
});
|
|
|
|
describe("GitHub Copilot Provider", () => {
|
|
it.skipIf(!githubCopilotToken)("OpenAI path should expose responseId", { retry: 3, timeout: 30000 }, async () => {
|
|
const llm = getModel("github-copilot", "gpt-5.3-codex");
|
|
await expectResponseId(llm, { apiKey: githubCopilotToken });
|
|
});
|
|
|
|
it.skipIf(!githubCopilotToken)(
|
|
"Anthropic path should expose responseId",
|
|
{ retry: 3, timeout: 30000 },
|
|
async () => {
|
|
const llm = getModel("github-copilot", "claude-sonnet-4");
|
|
await expectResponseId(llm, { apiKey: githubCopilotToken });
|
|
},
|
|
);
|
|
});
|
|
|
|
describe("Google Gemini CLI Provider", () => {
|
|
it.skipIf(!geminiCliToken)("should expose responseId", { retry: 3, timeout: 30000 }, async () => {
|
|
const llm = getModel("google-gemini-cli", "gemini-2.5-flash");
|
|
await expectResponseId(llm, { apiKey: geminiCliToken });
|
|
});
|
|
});
|
|
|
|
describe("Google Antigravity Provider", () => {
|
|
it.skipIf(!antigravityToken)("Gemini path should expose responseId", { retry: 3, timeout: 30000 }, async () => {
|
|
const llm = getModel("google-antigravity", "gemini-3.1-pro-high");
|
|
await expectResponseId(llm, { apiKey: antigravityToken });
|
|
});
|
|
|
|
it.skipIf(!antigravityToken)("Claude path should expose responseId", { retry: 3, timeout: 30000 }, async () => {
|
|
const llm = getModel("google-antigravity", "claude-sonnet-4-6");
|
|
await expectResponseId(llm, { apiKey: antigravityToken });
|
|
});
|
|
});
|
|
|
|
describe("OpenAI Codex Provider", () => {
|
|
it.skipIf(!openaiCodexToken)("should expose responseId", { retry: 3, timeout: 30000 }, async () => {
|
|
const llm = getModel("openai-codex", "gpt-5.2-codex");
|
|
await expectResponseId(llm, { apiKey: openaiCodexToken });
|
|
});
|
|
});
|
|
});
|