Alphanso Technology's private Ollama inference server — run powerful open-source LLMs with zero data leaving your environment.
Alphanso's Ollama server gives your team private, fast, and compliant access to open-source LLMs.
base_url to this server and you're live instantly.Pre-loaded open-source models available on this server. Pull additional models anytime.
Use the Ollama REST API or any OpenAI-compatible SDK to start generating.
# Generate a completion curl https://ollama.alphansolabs.in/api/generate \ -d '{ "model": "llama3.2", "prompt": "Hello from Alphanso!", "stream": false }'
from openai import OpenAI client = OpenAI( base_url="https://ollama.alphansolabs.in/v1", api_key="ollama" # not required ) res = client.chat.completions.create( model="llama3.2", messages=[ {"role": "user", "content": "What is AI?"} ] ) print(res.choices[0].message.content)
const res = await fetch( "https://ollama.alphansolabs.in/api/chat", { method: "POST", headers: { "Content-Type": "application/json" }, body: JSON.stringify({ model: "mistral", messages: [ { role: "user", content: "Hello!" } ] }) } ); const data = await res.json();
# List all available models curl https://ollama.alphansolabs.in/api/tags # Pull a new model curl https://ollama.alphansolabs.in/api/pull \ -d '{"name": "deepseek-r1:8b"}' # Check server health curl https://ollama.alphansolabs.in/api/version