Try agent mode in VS Code!
Dismiss this update
AI Toolkit supports a broad range of generative AI models. Both Small Language Models (SLM) and Large Language Models (LLM) are supported.
In the model catalog, you can access models from various sources:
To find a model in the model catalog:
Select the AI Toolkit view in the Activity Bar
Select CATALOG > Models to open the model catalog
Select a model card in the model catalog to view more details of the selected model.
Use the filters to reduce the list of available models
Hosted by: AI Toolkit supports GitHub, ONNX, OpenAI, Anthropic, Google as model hosting sources.
Publisher: The publisher for AI models, such as Microsoft, Meta, Google, OpenAI, Anthropic, Mistral AI, and more.
Tasks: Currently, only Text Generation
is supported.
Model type: Filter models that can run remotely or locally on CPU, GPU, or NPU. This filter depends on the local availability.
Fine-tuning Support: Show models that can be used to run fine-tuning.
To reference a self-hosted model or locally-running Ollama model:
Select + Add model in the model catalog
Choose between Ollama or a custom model in the model Quick Pick
Provide details to add the model
Some models require a publisher or hosting-service license and account to sign-in. In that case, before you can run the model in the model playground, you are prompted to provide this information.
AI Toolkit enables you to test run a model in the playground for chat completions. You have different options, available through the actions on the model card in the model catalog.
AI Toolkit's playground also supports remote models. If you have a self-hosted or deployed model that is accessible from the internet, you can add it to AI Toolkit and use it in the playground.
+
icon to add a remote model into AI Toolkit.Ollama enables many popular genAI models to run locally with CPU via GGUF quantization. If you have Ollama installed on your local machine with downloaded Ollama models, you can add them to AI Toolkit for use in the model playground.
Hover over MY MODELS in the tree view and select the "+" icon to add a model
Alternatively, select the + Add model button in the model catalog or playground.
Select Add an Ollama model
Next, select Select models from Ollama library
If you start the Ollama runtime at a different endpoint, choose Provide custom Ollama endpoint to specify an Ollama endpoint.
Select the models you want to add to AI Toolkit, and then select OK
AI Toolkit only shows models that are already downloaded in Ollama and not yet added to AI Toolkit. To download a model from Ollama, you can run ollama pull
. To see the list of models supported by Ollama, see the Ollama library or refer to the Ollama documentation.
You should now see the selected Ollama model(s) in the list of models in the tree view.
Attachment is not support yet for Ollama models. Since we connect to Ollama using its OpenAI compatible endpoint and it doesn't support attachments yet.