Model access

Not all models are freely accessible, so keep this in mind when setting up any LLM provider.

NOTE: If you have not been granted access, you’ll probably see an error like this:

InvalidRequestError(message='The model: `gpt-4` does not exist', param=None, code='model_not_found', http_status=404, request_id=None)

There is nothing this project can do to fix the error for you – contact OpenAI and request GPT-4 access.

Follow one of the methods below to utilize non-default models:

Method 1: Set a default preset configured with GPT-4o

SeePresets to configure a preset using GPT-4o

Add the preset to the config file as the default preset on startup:

# This assumes you created a preset named 'gpt-4o'model:default_preset:gpt-4o

Method 2: Dynamically switch

From within the shell, execute this command:

/model model_name gpt-4o

…or… if you’re not currently using the ‘chat_openai’ provider:

/provider chat_openai gpt-4o