Can you add Context Length to OpenAI Alt?

1. The problem

Ollama has no easy way to set context window for Local LLMs other than recreating the entire model with a new modelfile. So now we need a new LLM for each context size. So Silly.

2. Simple solution

Allow us to set Context Length along with Temperature when configuring an OpenAI Alt … I believe the param name is “num_ctx”

I had a search of num_ctx

Looks like there is a bug or missing feature set with ollama

I think num_ctx cant be used by openai api and only with direct calls to ollama using its own api. Ie you might have to use webhooks.

Otherwise is saw this.

There is open req to have openai max_token param map automatically to ollama num_ctx