Networking Issues
Configure Certificates
If you’re seeing afetch failed
error and your network requires custom certificates, you will need to configure them in your config file. In each of the objects in the "models"
array, add requestOptions.caBundlePath
like this:
requestOptions.caBundlePath
to an array of paths to multiple certificates.
Windows VS Code Users: Installing the win-ca extension should also correct this issue.
VS Code Proxy Settings
If you are using VS Code and require requests to be made through a proxy, you are likely already set up through VS Code’s Proxy Server Support. To double-check that this is enabled, usecmd/ctrl
+ ,
to open settings and search for “Proxy Support”. Unless it is set to “off”, then VS Code is responsible for making the request to the proxy.
code-server
Continue can be used in code-server, but if you are running across an error in the logs that includes “This is likely because the editor is not running in a secure context”, please see their documentation on securely exposing code-server.Changes to assistants not showing in VS Code
If you’ve made changes to assistants (adding, modifying, or removing them) but the changes aren’t appearing in the Continue extension in VS Code, try reloading the VS Code window:- Open the command palette (
cmd/ctrl
+shift
+P
) - Type “Reload Window”
- Select the reload option
I installed Continue, but don’t see the sidebar window
By default the Continue window is on the left side of VS Code, but it can be dragged to right side as well, which we recommend in our tutorial. In the situation where you have previously installed Continue and moved it to the right side, it may still be there. You can reveal Continue either by using cmd/ctrl+L or by clicking the button in the top right of VS Code to open the right sidebar.I’m getting a 404 error from OpenAI
If you have entered a valid API key and model, but are still getting a 404 error from OpenAI, this may be because you need to add credits to your billing account. You can do so from the billing console. If you just want to check that this is in fact the cause of the error, you can try adding $1 to your account and checking whether the error persists.I’m getting a 404 error from OpenRouter
If you have entered a valid API key and model, but are still getting a 404 error from OpenRouter, this may be because models that do not support function calling will return an error to Continue when a request is sent. Example error:HTTP 404 Not Found from https://openrouter.ai/api/v1/chat/completions
Indexing issues
If you are having persistent errors with indexing, our recommendation is to rebuild your index from scratch. Note that for large codebases this may take some time. This can be accomplished using the following command:Continue: Rebuild codebase index
.
Agent mode is unavailable or tools aren’t working
If Agent mode is grayed out or tools aren’t functioning properly, this is likely due to model capability configuration issues.Continue uses system message tools as a fallback for models without native tool support, so most models should work with Agent mode automatically.
Check if your model has tool support
- Not all models support native tool/function calling, but Continue will automatically use system message tools as a fallback
- Try adding
capabilities: ["tool_use"]
to your model config to force tool support - Verify your provider supports function calling or that system message tools are working correctly
Tools Not Working
If tools aren’t being called:- Ensure
tool_use
is in your capabilities - Check that your API endpoint actually supports function calling
- Some providers may use different function calling formats
Images Not Uploading
If you can’t upload images:- Add
image_input
to capabilities - Ensure your model actually supports vision (e.g., gpt-4-vision, claude-3)
- Check that your provider passes through image data
Add capabilities
If Continue’s autodetection isn’t working correctly, you can manually add capabilities in yourconfig.yaml
:
Verify with provider
Some proxy services (like OpenRouter) or custom deployments may not preserve tool calling capabilities. Check your provider’s documentation.Verifying Current Capabilities
To see what capabilities Continue detected for your model:- Check the mode selector tooltips - they indicate if tools are available
- Try uploading an image - if disabled, the model lacks
image_input
- Check if Agent mode is available - requires
tool_use
Android Studio - “Nothing to show” in Chat
This can be fixed by selectingActions > Choose Boot runtime for the IDE
then selecting the latest version, and then restarting Android Studio. See this thread for details.
I received a “Codebase indexing disabled - Your Linux system lacks required CPU features (AVX2, FMA)” notification
We use LanceDB as our vector database for codebase search features. On x64 Linux systems, LanceDB requires specific CPU features (FMA and AVX2) which may not be available on older processors. Most Continue features will work normally, including autocomplete and chat. However, commands that rely on codebase indexing, such as@codebase
, @files
, and @folder
, will be disabled.
For more details about this requirement, see the LanceDB issue #2195.
Ollama Issues
For a comprehensive guide on setting up and troubleshooting Ollama, see the Ollama Guide.Unable to connect to local Ollama instance
If you’re getting “Unable to connect to local Ollama instance” errors:- Verify Ollama is running: Check http://localhost:11434 in your browser - you should see “Ollama is running”
- Start Ollama properly: Use
ollama serve
(not justollama run model-name
) - Check your config: Ensure your
config.yaml
has the correct setup:
Connection failed to remote Ollama (EHOSTUNREACH/ECONNREFUSED)
When connecting to Ollama on another machine:-
Configure Ollama to listen on all interfaces:
- Set environment variable:
OLLAMA_HOST=0.0.0.0:11434
- For systemd: Edit
/etc/systemd/system/ollama.service
and add under[Service]
: - Restart Ollama:
sudo systemctl restart ollama
- Set environment variable:
- Update your Continue config:
- Check firewall settings: Ensure port 11434 is open on the server
Ollama not working in WSL
For WSL users having connection issues:Windows 11 22H2+ (Recommended)
Create or edit%UserProfile%\.wslconfig
:
wsl --shutdown
Older Windows/WSL versions
In PowerShell (as Administrator):Docker container can’t connect to host Ollama
When running Continue or other tools in Docker that need to connect to Ollama on the host: Windows/Mac: Usehost.docker.internal
:
172.17.0.1
):
Parse errors with remote Ollama
If you’re getting parse errors with remote Ollama:-
Verify the model is installed on the remote:
-
Install missing models:
-
Check URL format: Ensure you’re using
http://
nothttps://
for local network addresses
Local Assistant
Managing Local Secrets and Environment Variables
For running Continue completely offline without internet access, see the Running Continue Without Internet guide. Continue supports multiple methods for managing secrets locally, searched in this order:- Workspace
.env
files: Place a.env
file in your workspace root directory - Workspace Continue folder: Place a
.env
file in<workspace-root>/.continue/.env
- Global
.env
file: Place a.env
file in~/.continue/.env
for user-wide secrets - Process environment variables: Use standard system environment variables
Creating .env
files
Create a .env
file in one of these locations:
- Per-workspace:
<workspace-root>/.env
or<workspace-root>/.continue/.env
- Global:
~/.continue/.env
.env
file:
Using secrets in config.yaml
Reference your local secrets using thesecrets
namespace:
Hub-managed secrets
For centralized team secret management, use${{ inputs.SECRET_NAME }}
syntax in your config.yaml and manage them at https://hub.continue.dev/settings/secrets:
Important notes
- Never commit
.env
files to version control - add them to.gitignore
- The
.env
file uses standard dotenv format (KEY=value, no quotes needed) - Secrets are loaded when Continue starts, so restart your IDE after changes
- Local
.env
files take precedence over Hub secrets when both exist
Troubleshooting secrets
If your API keys aren’t being recognized:- Check the
.env
file is in the correct location - Ensure there are no quotes around values in the
.env
file - Restart your IDE after adding/changing secrets
- Verify the variable name matches exactly (case-sensitive)
- Check that your
.env
file has proper line endings (LF, not CRLF on Windows)
Using Model Addons Locally
You can leverage model addons from the Continue Hub in your local assistant configurations using theuses:
syntax. This allows you to reference pre-configured model blocks without duplicating configuration.
Requirements
- You must be logged in to Continue
- Internet connection is required (model addons are fetched from the hub)
Usage
In your localconfig.yaml
, reference model addons using the format provider/model-name
:
With local configuration
You can combine hub model addons with local models:Override addon settings
You can override specific settings from the model addon:How do I reset the state of the extension?
Continue stores its data in the~/.continue
directory (%USERPROFILE%\.continue
on Windows).
If you’d like to perform a clean reset of the extension, including removing all configuration files, indices, etc, you can remove this directory, uninstall, and then reinstall.