-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Create an appliance for ray framework #157
Labels
category: appliances
Generic category for new appliances
status: accepted
The issue is valid and will be planned for fix or implementation
type: feature
A new feature planned for implementation
Milestone
Comments
rsmontero
pushed a commit
that referenced
this issue
Dec 20, 2024
- ONE_APP_RAY_MODEL_URL or ONE_APP_RAY_MODEL64: the python model file url for download it or the file content encoded in base64, respectively. - ONE_APP_RAY_CONFIG_FILE or ONE_APP_RAY_CONFIG64 will contain the Ray configuration yaml file content (in yaml plaintext or encoded in base64). If not provided, the config.yaml.templatefile will be parameterized with the following attributes: - ONE_APP_RAY_SERVE_PORT: The port where the chatbot API will be served. Defaults to '8000' - ONE_APP_RAY_MODEL_ID = The language model id to use for the chatbot app. Defaults to meta-llama/Llama-3.2-1B - ONE_APP_RAY_TOKEN: The token used for calling the model - ONE_APP_RAY_TEMPERATURE: String with language model temperature float number (from 0.0 to 1.0). Defaults to 0.1 - ONE_APP_RAY_CHATBOT_REPLICAS: num of chatbot deployed replicas - ONE_APP_RAY_CHATBOT_CPUS: chatbot app allocated CPU Signed-off-by: Aleix Ramírez <[email protected]>
rsmontero
added a commit
that referenced
this issue
Dec 20, 2024
- Linting - Update vatiable naming - Update VM with chatbot endpoint - Add default model.py Template configuration - DISK should be no less than 15Gb - Memory should be 8Gb or greater - User inputs USER_INPUTS=[ ONEAPP_RAY_API_PORT="O|number|port for api server| |8000", ONEAPP_RAY_CONFIG_FILE64="O|text|Configuration file raw| |", ONEAPP_RAY_MODEL_FILE64="O|text64|Python script to use as model| |", ONEAPP_RAY_MODEL_ID="O|list-multiple|AI model to use|meta-llama/Llama-3.2-1B-Instruct,Qwen/Qwen2.5-1.5B-Instruct|", ONEAPP_RAY_MODEL_TEMPERATURE="O|number-float|| |0.1", ONEAPP_RAY_MODEL_TOKEN="O|text|token for the model| | ******", ONEAPP_RAY_MODEL_URL="O|text|URL to download de model| |" ] - Context CONTEXT=[ NETWORK="YES", ONEAPP_RAY_API_PORT="$ONEAPP_RAY_API_PORT", ONEAPP_RAY_CHATBOT_CPUS="$VCPU", ONEAPP_RAY_CONFIG_FILE64="$ONEAPP_RAY_CONFIG_FILE64", ONEAPP_RAY_MODEL_FILE64="$ONEAPP_RAY_MODEL_FILE64", ONEAPP_RAY_MODEL_ID="$ONEAPP_RAY_MODEL_ID", ONEAPP_RAY_MODEL_TEMPERATURE="$ONEAPP_RAY_MODEL_TEMPERATURE", ONEAPP_RAY_MODEL_TOKEN="$ONEAPP_RAY_MODEL_TOKEN", ONEAPP_RAY_MODEL_URL="$ONEAPP_RAY_MODEL_URL", SET_HOSTNAME="$NAME", SSH_PUBLIC_KEY="$USER[SSH_PUBLIC_KEY]", TOKEN="YES" ]
rsmontero
added a commit
that referenced
this issue
Dec 26, 2024
Moves some of the configuration options under its own section (APPLICATION). Template configuration: CONTEXT=[ NETWORK="YES", ONEAPP_RAY_API_PORT="$ONEAPP_RAY_API_PORT", ONEAPP_RAY_API_ROUTE="$ONEAPP_RAY_API_ROUTE", ONEAPP_RAY_APPLICATION_FILE64="$ONEAPP_RAY_MODEL_FILE64", ONEAPP_RAY_APPLICATION_URL="$ONEAPP_RAY_MODEL_URL", ONEAPP_RAY_CHATBOT_CPUS="$VCPU", ONEAPP_RAY_MODEL_ID="$ONEAPP_RAY_MODEL_ID", ONEAPP_RAY_MODEL_TEMPERATURE="$ONEAPP_RAY_MODEL_TEMPERATURE", ONEAPP_RAY_MODEL_TOKEN="$ONEAPP_RAY_MODEL_TOKEN", SET_HOSTNAME="$NAME", SSH_PUBLIC_KEY="$USER[SSH_PUBLIC_KEY]", TOKEN="YES" ] INPUTS_ORDER="ONEAPP_RAY_API_PORT,ONEAPP_RAY_MODEL_ID,ONEAPP_RAY_MODEL_TEMPERATURE,ONEAPP_RAY_MODEL_TOKEN" USER_INPUTS=[ ONEAPP_RAY_API_PORT="O|number|Port number for the API endpoint.| |8000", ONEAPP_RAY_API_ROUTE="O|text|Route path for the REST API exposed by the RAY application.| |/chat", ONEAPP_RAY_APPLICATION_FILE64="O|text64|Python application to be deployed in the RAY framework (base64).| |", ONEAPP_RAY_APPLICATION_URL="O|text|URL to download the Python application for the RAY framework.| |", ONEAPP_RAY_MODEL_ID="M|list-multiple|Determines the AI model(s) to use for inference.|meta-llama/Llama-3.2-1B-Instruct,Qwen/Qwen2.5-1.5B-Instruct|", ONEAPP_RAY_MODEL_TEMPERATURE="M|number-float|Temperature parameter for model outputs, controlling the randomness of generated text.| |0.1", ONEAPP_RAY_MODEL_TOKEN="M|text|Provides the authentication token required to access the specified AI model. | |********" ]
rsmontero
added a commit
that referenced
this issue
Dec 31, 2024
rsmontero
added a commit
that referenced
this issue
Jan 2, 2025
rsmontero
added a commit
to OpenNebula/marketplace
that referenced
this issue
Jan 3, 2025
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Labels
category: appliances
Generic category for new appliances
status: accepted
The issue is valid and will be planned for fix or implementation
type: feature
A new feature planned for implementation
This issue is to create an appliance with pre-installed ray
The text was updated successfully, but these errors were encountered: