diff --git a/README.md b/README.md index 590a69e..d5def39 100644 --- a/README.md +++ b/README.md @@ -177,7 +177,7 @@ The following example showcases how you can integrate GenAI into your project. import { GenAIModel } from '@ibm-generative-ai/node-sdk/langchain'; const model = new GenAIModel({ - modelId: 'google/ul2', + modelId: 'google/flan-ul2', parameters: {}, configuration: { apiKey: 'pak-.....', @@ -222,7 +222,7 @@ console.log(text); // ArcticAegis import { GenAIModel } from '@ibm-generative-ai/node-sdk/langchain'; const model = new GenAIModel({ - modelId: 'google/ul2', + modelId: 'google/flan-ul2', stream: true, parameters: {}, configuration: { @@ -245,7 +245,7 @@ await model.call('Tell me a joke.', undefined, [ import { GenAIChatModel } from '@ibm-generative-ai/node-sdk/langchain'; const client = new GenAIChatModel({ - modelId: 'togethercomputer/gpt-neoxt-chat-base-20b', + modelId: 'eleutherai/gpt-neox-20b', stream: false, configuration: { endpoint: process.env.ENDPOINT, diff --git a/examples/chat.ts b/examples/chat.ts index 1217790..454d3f7 100644 --- a/examples/chat.ts +++ b/examples/chat.ts @@ -4,7 +4,7 @@ const client = new Client({ apiKey: process.env.GENAI_API_KEY, }); -const model_id = 'google/ul2'; +const model_id = 'google/flan-ul2'; { // Start a conversation @@ -69,7 +69,7 @@ const model_id = 'google/ul2'; // Streaming callbacks client.chat( { - model_id: 'google/ul2', + model_id: 'google/flan-ul2', messages: [{ role: 'user', content: 'How are you?' }], }, { stream: true }, diff --git a/examples/langchain/llm-chat.ts b/examples/langchain/llm-chat.ts index 6f0921b..1e5ebe5 100644 --- a/examples/langchain/llm-chat.ts +++ b/examples/langchain/llm-chat.ts @@ -4,7 +4,7 @@ import { GenAIChatModel } from '../../src/langchain/llm-chat.js'; const makeClient = (stream?: boolean) => new GenAIChatModel({ - modelId: 'togethercomputer/gpt-neoxt-chat-base-20b', + modelId: 'eleutherai/gpt-neox-20b', stream, configuration: { endpoint: process.env.ENDPOINT, diff --git a/src/tests/e2e/client.test.ts b/src/tests/e2e/client.test.ts index 44637b6..90d403c 100644 --- a/src/tests/e2e/client.test.ts +++ b/src/tests/e2e/client.test.ts @@ -58,7 +58,7 @@ describe('client', () => { const makeValidStream = (parameters: Record = {}) => client.generate( { - model_id: 'google/ul2', + model_id: 'google/flan-ul2', input: 'Hello, World', parameters: { max_new_tokens: 10, @@ -89,11 +89,11 @@ describe('client', () => { client.generate( [ { - model_id: 'google/ul2', + model_id: 'google/flan-ul2', input: 'Hello, World', }, { - model_id: 'google/ul2', + model_id: 'google/flan-ul2', input: 'Hello, World', }, ] as unknown as GenerateInput, @@ -153,7 +153,7 @@ describe('client', () => { const chunks: GenerateOutput[] = []; client.generate( { - model_id: 'google/ul2', + model_id: 'google/flan-ul2', input: 'Hello, World', parameters: {}, }, @@ -215,7 +215,7 @@ describe('client', () => { const makeValidStream = () => client.chat( { - model_id: 'google/ul2', + model_id: 'google/flan-ul2', messages: [{ role: 'user', content: 'Hello World!' }], }, { @@ -246,7 +246,7 @@ describe('client', () => { const chunks: ChatOutput[] = []; client.chat( { - model_id: 'google/ul2', + model_id: 'google/flan-ul2', messages: [{ role: 'user', content: 'Hello World!' }], }, { diff --git a/src/tests/e2e/langchain/llm-chat.test.ts b/src/tests/e2e/langchain/llm-chat.test.ts index 6add30f..2309f4c 100644 --- a/src/tests/e2e/langchain/llm-chat.test.ts +++ b/src/tests/e2e/langchain/llm-chat.test.ts @@ -9,7 +9,7 @@ describeIf(process.env.RUN_LANGCHAIN_CHAT_TESTS === 'true')( () => { const makeClient = (stream?: boolean) => new GenAIChatModel({ - modelId: 'togethercomputer/gpt-neoxt-chat-base-20b', + modelId: 'eleutherai/gpt-neox-20b', stream, configuration: { endpoint: process.env.ENDPOINT, diff --git a/src/tests/integration/client.test.ts b/src/tests/integration/client.test.ts index 93ec1b6..20d21f9 100644 --- a/src/tests/integration/client.test.ts +++ b/src/tests/integration/client.test.ts @@ -71,7 +71,7 @@ describe('client', () => { test('should replace the config', async () => { const input = { - model_id: 'google/ul2', + model_id: 'google/flan-ul2', parameters: { decoding_method: 'greedy', random_seed: 8, @@ -98,7 +98,7 @@ describe('client', () => { test('should set and reset the config', async () => { const input = { - model_id: 'google/ul2', + model_id: 'google/flan-ul2', parameters: { decoding_method: 'greedy', random_seed: 8,