Skip to content

Configuring OpenAI-Compatible Models

Configure OpenAI-compatible Fill In the Middle models in Flexpilot IDE using cloud providers like OpenAI or self-hosted solutions such as Ollama, Lm Studio, LocalAI, and vLLM. Gather the required details and follow the setup guide to configure the model.

Configuration Reference

Use the following settings to configure an OpenAI-compatible API provider in Flexpilot:

⚙️ Configuration Reference
ParameterPurposeSample Value
API Base URLPrimary endpointhttps://codestral.mistral.ai/v1/fim
API KeyAuthentication tokenGenerated from Setup Guide
Model IDAI model selectioncodegemma:2b
Max Output TokensResponse length limit500
Max Input TokensContext window size3000
TemperatureCreativity control (0-1)0.2 for precise code generation
Debounce WaitCompletion trigger delay200 milliseconds

💡 Pro Tips

  • Use lower temperatures (0.1-0.3) for more deterministic code generation.
  • For snappier completions, consider reducing Debounce Wait to 100ms.

Available Providers

Below is a list of some popular OpenAI-compatible API providers and services that you can use with Flexpilot, with the help of OpenAI Compatible Completion Model Provider.

NOTE

The following list is not exhaustive and may not include all available providers. Please refer to the provider's documentation for more details.

  • DeepSeek API
  • Ollama
  • Hugging Face
  • Replicate
  • Together AI
  • Fireworks AI

Next Steps

Set up the API credentials obtained from this page's instructions in the Flexpilot IDE to enable AI-powered code completions. For detailed guidance, refer to the Setup Guide.