Ollama allows you to run open-weight LLMs locally on your machine.
Setting up Ollama¶
Install Ollama following the instructions at https://
ollama .com /download Pull a model, for example:
ollama pull llama3.2Start the Ollama server (it typically runs on
http://localhost:11434)
Configuring jupyterlite-ai to use Ollama¶
In JupyterLab, open the AI settings panel and go to the Providers section
Click on “Add a new provider”
Select the Generic (OpenAI-compatible) provider
Configure the following settings:
Base URL: Select
http://localhost:11434/v1from the suggestions (or enter manually)Model: The model name you pulled (e.g.,
llama3.2)API Key: Leave empty (not required for Ollama)