diff --git a/fern/03-reference/baml/clients/providers/anthropic.mdx b/fern/03-reference/baml/clients/providers/anthropic.mdx index 0f81e5693..33cbdec65 100644 --- a/fern/03-reference/baml/clients/providers/anthropic.mdx +++ b/fern/03-reference/baml/clients/providers/anthropic.mdx @@ -67,6 +67,8 @@ client MyClient { ``` + + diff --git a/fern/03-reference/baml/clients/providers/aws-bedrock.mdx b/fern/03-reference/baml/clients/providers/aws-bedrock.mdx index f359f1a3d..32a29f4d9 100644 --- a/fern/03-reference/baml/clients/providers/aws-bedrock.mdx +++ b/fern/03-reference/baml/clients/providers/aws-bedrock.mdx @@ -13,8 +13,13 @@ Example: client MyClient { provider aws-bedrock options { - model_id "anthropic.claude-3-5-sonnet-20240620-v1:0" - api_key env.MY_OPENAI_KEY + inference_configuration { + max_tokens 100 + } + // model_id "mistral.mistral-7b-instruct-v0:2" + // model "anthropic.claude-3-5-sonnet-20240620-v1:0" + // model_id "anthropic.claude-3-haiku-20240307-v1:0" + model "meta.llama3-8b-instruct-v1:0" } } ``` @@ -26,9 +31,9 @@ mechanisms supported by the SDK](https://docs.rs/aws-config/latest/aws_config/index.html), including but not limited to: - - `AWS_ACCESS_KEY_ID` and `AWS_SECRET_ACCESS_KEY` as set in your environment variables - loading the specified `AWS_PROFILE` from `~/.aws/config` - built-in authn for services running in EC2, ECS, Lambda, etc. + - You can also specify the access key ID, secret access key, and region directly (see below) ## Playground setup @@ -50,6 +55,8 @@ Add these three environment variables to your extension variables to use the AWS We don't have any checks for this field, you can pass any string you wish. + + @@ -62,6 +69,21 @@ Add these three environment variables to your extension variables to use the AWS We don't have any checks for this field, you can pass any string you wish. + + The AWS access key ID to use. **Default: `AWS_ACCESS_KEY_ID` environment variable** + + + + The AWS secret access key to use. **Default: `AWS_SECRET_ACCESS_KEY` environment variable** + + + ## Forwarded options The model to use. @@ -110,7 +132,6 @@ client MyClient { max_tokens 1000 temperature 1.0 top_p 0.8 - stop_sequence ["_EOF"] } } } diff --git a/fern/03-reference/baml/clients/providers/azure.mdx b/fern/03-reference/baml/clients/providers/azure.mdx index d56047b7f..029bc8680 100644 --- a/fern/03-reference/baml/clients/providers/azure.mdx +++ b/fern/03-reference/baml/clients/providers/azure.mdx @@ -95,6 +95,8 @@ client MyClient { ``` + + diff --git a/fern/03-reference/baml/clients/providers/google-ai.mdx b/fern/03-reference/baml/clients/providers/google-ai.mdx index 52ed52106..df5a093d8 100644 --- a/fern/03-reference/baml/clients/providers/google-ai.mdx +++ b/fern/03-reference/baml/clients/providers/google-ai.mdx @@ -81,6 +81,8 @@ client MyClient { ``` + + diff --git a/fern/03-reference/baml/clients/providers/ollama.mdx b/fern/03-reference/baml/clients/providers/ollama.mdx index 99d3478f0..9821db860 100644 --- a/fern/03-reference/baml/clients/providers/ollama.mdx +++ b/fern/03-reference/baml/clients/providers/ollama.mdx @@ -61,6 +61,8 @@ client MyClient { ``` + + diff --git a/fern/03-reference/baml/clients/providers/openai-generic.mdx b/fern/03-reference/baml/clients/providers/openai-generic.mdx index 8042d42cc..a6bd929bb 100644 --- a/fern/03-reference/baml/clients/providers/openai-generic.mdx +++ b/fern/03-reference/baml/clients/providers/openai-generic.mdx @@ -62,6 +62,8 @@ client MyClient { + + diff --git a/fern/03-reference/baml/clients/providers/openai.mdx b/fern/03-reference/baml/clients/providers/openai.mdx index b1a32ac7a..9604bff1a 100644 --- a/fern/03-reference/baml/clients/providers/openai.mdx +++ b/fern/03-reference/baml/clients/providers/openai.mdx @@ -70,6 +70,8 @@ client MyClient { + + diff --git a/fern/03-reference/baml/clients/providers/vertex.mdx b/fern/03-reference/baml/clients/providers/vertex.mdx index a2067f833..b8496dc5d 100644 --- a/fern/03-reference/baml/clients/providers/vertex.mdx +++ b/fern/03-reference/baml/clients/providers/vertex.mdx @@ -171,15 +171,6 @@ The options are passed through directly to the API, barring a few. Here's a shor - - - The default role for any prompts that don't specify a role. **Default: `user`** - - - MyClient { ``` + + diff --git a/fern/snippets/role-selection.mdx b/fern/snippets/role-selection.mdx index d5957c59b..94eb165dd 100644 --- a/fern/snippets/role-selection.mdx +++ b/fern/snippets/role-selection.mdx @@ -13,5 +13,5 @@ > Which roles should we forward to the API? **Default: `["system", "user", "assistant"]` usually, but some models like OpenAI's `o1-mini` will use `["user", "assistant"]`** - Anything not in this list will be set to the `default_role`. + When building prompts, any role not in this list will be set to the `default_role`.