freshcrate
Home > Infrastructure > ollamafreeapi

ollamafreeapi

OllamaFreeAPI: Free Distributed API for Ollama LLMs Public gateway to our managed Ollama servers with: - Zero-configuration access to 50+ models - Auto load-balanced across global nodes - Free tier w

Description

OllamaFreeAPI: Free Distributed API for Ollama LLMs Public gateway to our managed Ollama servers with: - Zero-configuration access to 50+ models - Auto load-balanced across global nodes - Free tier with no API keys required - Built on ollama-python

README

OllamaFreeAPI

PyPI Version Python Versions License Free API

Unlock AI Innovation for Free

Access the world's best open language models in one place!

OllamaFreeAPI provides free access to leading open-source LLMs including:

  • πŸ¦™ LLaMA (Meta)
  • πŸŒͺ️ Mistral (Mistral AI)
  • πŸ” DeepSeek (DeepSeek)
  • πŸ¦„ Qwen (Alibaba Cloud)

No payments. No credit cards. Just pure AI power at your fingertips.

# Install or upgrade to the latest version
pip install ollamafreeapi --upgrade

πŸ“š Documentation

Why Choose OllamaFreeAPI?

Feature Others OllamaFreeAPI
Free Access ❌ Limited trials βœ… Always free
Model Variety 3-5 models Verified endpoints only
Reliability Highly variable Validated active models
Ease of Use Complex setup Zero-config
Community Support Paid only Free & active

πŸ“Š Project Statistics

Here are some key statistics about the current state of OllamaFreeAPI:

  • Active Models: 16 (Ready to use and tested)
  • Model Families: 3 (gemma, llama, qwen)
  • Endpoints: 6 highly reliable server nodes

πŸš€ Quick Start

Streaming Example

from ollamafreeapi import OllamaFreeAPI

client = OllamaFreeAPI()

# Stream responses in real-time
for chunk in client.stream_chat('What is quantum computing?', model='llama3.2:3b'):
    print(chunk, end='', flush=True)

Non-Streaming Example

from ollamafreeapi import OllamaFreeAPI

client = OllamaFreeAPI()

# Get instant responses
response = client.chat(
    model="gpt-oss:20b",
    prompt="Explain neural networks like I'm five",
    temperature=0.7
)
print(response)

🌟 Featured Models

Popular Foundation Models

  • llama3.2:3b - Meta's efficient 3.2B parameter model
  • deepseek-r1:latest - Strong reasoning capabilities built on Qwen
  • gpt-oss:20b - Powerful Gemma-based 20B completion model
  • mistral:latest - High-performance baseline Mistral model

Specialized Models

  • mistral-nemo:custom - 12.2B open weights language model
  • bakllava:latest - Vision and language model
  • smollm2:135m - Extremely lightweight assistant

🌍 Global Infrastructure

Our free API is powered by distributed community nodes:

  • Fast response times
  • Automatic load balancing and server selection
  • Real-time availability checks

πŸ“„ API Reference

Core Methods

# List available models
api.list_models()  

# Get model details
api.get_model_info("mistral:latest")  

# Generate text
api.chat(model="llama3.2:3b", prompt="Your message")

# Stream responses
for chunk in api.stream_chat(prompt="Hello!", model="llama3:latest"):
    print(chunk, end='')

Advanced Features

# Check server locations
api.get_model_servers("deepseek-r1:latest")

# Generate raw API request
api.generate_api_request(model="llama3.2:3b", prompt="Hello")

# Get random model parameters (useful for LangChain integration)
api.get_llm_params()

🀝 Contributing

We welcome contributions! Please see our Contributing Guide for details.

πŸ“„ License

Open-source MIT license - View License

πŸ”— Links

Release History

VersionChangesUrgencyDate
main@2026-04-15Latest activity on main branchHigh4/15/2026
v0.1.4Latest release: v0.1.4Medium3/24/2026

Dependencies & License Audit

Loading dependencies...

Similar Packages

restaiRESTai is an AIaaS (AI as a Service) open-source platform. Supports many public and local LLM suported by Ollama/vLLM/etc. Precise embeddings usage, tuning, analytics etc. Built-in image/audio generatv6.1.45
SmarterRouterSmarterRouter: An intelligent LLM gateway and VRAM-aware router for Ollama, llama.cpp, and OpenAI. Features semantic caching, model profiling, and automatic failover for local AI labs.2.2.5
UltraRAGA Low-Code MCP Framework for Building Complex and Innovative RAG Pipelinesv0.3.0.2
lm-proxyOpenAI-compatible HTTP LLM proxy / gateway for multi-provider inference (Google, Anthropic, OpenAI, PyTorch). Lightweight, extensible Python/FastAPIβ€”use as library or standalone service.v3.2.2
KawaiiGPTKawaiiGPT β€” Open-source LLM gateway accessing DeepSeek, Gemini, and Kimi-K2 through reverse-engineered Pollinations API with no API keys required, built-in prompt injection capabilities for security rKawaiiGPT