This is generated documentation for the llama2 component.
Name | Type | Description |
---|---|---|
model_dir |
string | |
model |
string | |
tokenizer |
string |
Name | Type | Description |
---|---|---|
max_length |
uint? | |
temperature |
float? | |
top_p |
float? | |
repeat_penalty |
float? |
Name | Type | Description |
---|---|---|
prompt |
string |
Name | Type | Description |
---|---|---|
output |
string |
Given the following configuration:
Component-wide configuration as config.json
:
{
"model_dir": "XXX",
"model": "XXX",
"tokenizer": "XXX"
}
Operation configuration as op-config.json
:
{
"max_length": "XXX",
"temperature": "XXX",
"top_p": "XXX",
"repeat_penalty": "XXX"
}
$ wick invoke candle_ml/llama2:0.0.1 generate [email protected] [email protected] -- --prompt="XXX"
Or with inline configuration:
$ wick invoke candle_ml/llama2:0.0.1 generate \
--with='{ "model_dir":"XXX","model":"XXX","tokenizer":"XXX" }' \
--op-with='{ "max_length":"XXX","temperature":"XXX","top_p":"XXX","repeat_penalty":"XXX" }' \
-- --prompt="XXX"