Quickstart
Learn how to create, connect to, and delete a Managed Inference endpoint in a few steps.
View QuickstartEffortlessly deploy AI models on a sovereign infrastructure, manage and scale inference with full data privacy. Start now with a simple interface for creating dedicated endpoints.
Managed Inference QuickstartLearn how to create, connect to, and delete a Managed Inference endpoint in a few steps.
View QuickstartCore concepts that give you a better understanding of Scaleway Managed Inference.
View ConceptsCheck our guides about creating and managing Managed Inference endpoints.
View How-tosGuides to help you choose a Managed Inference endpoint, understand pricing and advanced configuration.
View additional contentLearn how to create and manage your Scaleway Managed Inference endpoints through the API.
Go to Managed Inference APILlama 3.1 Nemotron 70B and Molmo 72B are available for deployment on Managed Inference.
Nemotron improves human-like responses in complex tasks, while Molmo provides increased accuracy on multimodal inputs (text and images).
Managed Inference deployments can be created and managed with Infrastructure as Code using Scaleway's Terraform provider.
Find example usage and reference for the new scaleway_inference_deployment
resource in the official documentation.
Function calling allows a large language model (LLM) to interact with external tools or APIs.
Parameters tools
and tool_choice
of our OpenAI-compatible chat API are now accepted for models with this capacity.
Read our dedicated documentation and tutorial to get started!
Visit our Help Center and find the answers to your most frequent questions.
Visit Help Center