Elisp library for the OpenAI API
The OpenAI Elisp library provides convenient access to the OpenAI API from applications written in the Elips language.
P.S. This package is expected to be used as a library, so there are only a few interactable commands you can use, and those are mostly examples.
Table of Contents
You will need to set up your API key before you can use this library.
(setq openai-key "[YOUR API KEY]")
For requests that need your user identifier,
(setq openai-user "[YOUR USER UID]")
💡 Tip
The two variables
openai-key
andopenai-user
are the default values for sending requests! However, you can still overwrite the value by passing the keywords:key
and:user
!
Here is the simplest example that teaches you how to use this library. This is
a function with a query
and a callback function.
(openai-completion "How are you?"
(lambda (data)
(message "%s" data)))
All arguments are exposed in the argument list, so you can send any request in any way you want.
For example, the request function openai-completion
accepts argument
max-tokens
. By seeing OpenAI's references page:
max_tokens
integer Optional Defaults to 16The maximum number of tokens to generate in the completion.
The token count of your prompt plus
max_tokens
cannot exceed the model's context length. Most models have a context length of 2048 tokens (except for the newest models, which support 4096).
(openai-completion ...
...
:max-tokens 4069) ; max out tokens!
The API functions are followed by this pattern:
[PACKAGE NAME]-[API TYPE]-[REQUEST NAME]
For example:
(openai-file-list ...)
openai
- is the package namefile
- is the api type, see OpenAI API referencelist
- is the request name
The function's parameters are followed in this order:
- required - variables are required for this type of request
callback
- execution after the request is made- optional - other variables that are not required, but will affect the final output
(openai-completion "How are you?" ; required
(lambda (data) ; callback
...)
:max-tokens 4069) ; optional
Every type of request has a default model
, and we hope this benefits the users
to not worry about what model to use for their request! However, if you want to
use other models, you can use the keyword :model
to replace them!
(openai-completion ...
...
:model "text-davinci-003") ; replace the default model
While playing through this library, you might see this error quite often.
400 - Bad request. Please check error message and your parameters
Try set the variable openai--show-log
to t
, it will show more error messages.
If you would like to contribute to this project, you may either clone and make pull requests to this repository. Or you can clone the project and establish your own branch of this tool. Any methods are welcome!