freshcrate
Home > Infrastructure > new-api

new-api

A unified AI model hub for aggregation & distribution. It supports cross-converting various LLMs into OpenAI-compatible, Claude-compatible, or Gemini-compatible formats. A centralized gateway for pers

Description

A unified AI model hub for aggregation & distribution. It supports cross-converting various LLMs into OpenAI-compatible, Claude-compatible, or Gemini-compatible formats. A centralized gateway for personal and enterprise model management. ๐Ÿฅ

README

new-api

New API

๐Ÿฅ Next-Generation LLM Gateway and AI Asset Management System

็ฎ€ไฝ“ไธญๆ–‡ | ็น้ซ”ไธญๆ–‡ | English | Franรงais | ๆ—ฅๆœฌ่ชž

license release docker GoReportCard

QuantumNous%2Fnew-api | Trendshift
Featured๏ฝœHelloGitHub New API - All-in-one AI asset management gateway. | Product Hunt

Quick Start โ€ข Key Features โ€ข Deployment โ€ข Documentation โ€ข Help

๐Ÿ“ Project Description

Important


๐Ÿค Trusted Partners

No particular order

Cherry Studio Aion UI Peking University UCloud Alibaba Cloud IO.NET


๐Ÿ™ Special Thanks

JetBrains Logo

Thanks to JetBrains for providing free open-source development license for this project


๐Ÿš€ Quick Start

Using Docker Compose (Recommended)

# Clone the project
git clone https://github.com/QuantumNous/new-api.git
cd new-api

# Edit docker-compose.yml configuration
nano docker-compose.yml

# Start the service
docker-compose up -d
Using Docker Commands
# Pull the latest image
docker pull calciumion/new-api:latest

# Using SQLite (default)
docker run --name new-api -d --restart always \
  -p 3000:3000 \
  -e TZ=Asia/Shanghai \
  -v ./data:/data \
  calciumion/new-api:latest

# Using MySQL
docker run --name new-api -d --restart always \
  -p 3000:3000 \
  -e SQL_DSN="root:123456@tcp(localhost:3306)/oneapi" \
  -e TZ=Asia/Shanghai \
  -v ./data:/data \
  calciumion/new-api:latest

๐Ÿ’ก Tip: -v ./data:/data will save data in the data folder of the current directory, you can also change it to an absolute path like -v /your/custom/path:/data


๐ŸŽ‰ After deployment is complete, visit http://localhost:3000 to start using!

๐Ÿ“– For more deployment methods, please refer to Deployment Guide


๐Ÿ“š Documentation

Quick Navigation:

Category Link
๐Ÿš€ Deployment Guide Installation Documentation
โš™๏ธ Environment Configuration Environment Variables
๐Ÿ“ก API Documentation API Documentation
โ“ FAQ FAQ
๐Ÿ’ฌ Community Interaction Communication Channels

โœจ Key Features

For detailed features, please refer to Features Introduction

๐ŸŽจ Core Functions

Feature Description
๐ŸŽจ New UI Modern user interface design
๐ŸŒ Multi-language Supports Simplified Chinese, Traditional Chinese, English, French, Japanese
๐Ÿ”„ Data Compatibility Fully compatible with the original One API database
๐Ÿ“ˆ Data Dashboard Visual console and statistical analysis
๐Ÿ”’ Permission Management Token grouping, model restrictions, user management

๐Ÿ’ฐ Payment and Billing

  • โœ… Online recharge (EPay, Stripe)
  • โœ… Pay-per-use model pricing
  • โœ… Cache billing support (OpenAI, Azure, DeepSeek, Claude, Qwen and all supported models)
  • โœ… Flexible billing policy configuration

๐Ÿ” Authorization and Security

  • ๐Ÿ˜ˆ Discord authorization login
  • ๐Ÿค– LinuxDO authorization login
  • ๐Ÿ“ฑ Telegram authorization login
  • ๐Ÿ”‘ OIDC unified authentication
  • ๐Ÿ” Key quota query usage (with neko-api-key-tool)

๐Ÿš€ Advanced Features

API Format Support:

Intelligent Routing:

  • โš–๏ธ Channel weighted random
  • ๐Ÿ”„ Automatic retry on failure
  • ๐Ÿšฆ User-level model rate limiting

Format Conversion:

  • ๐Ÿ”„ OpenAI Compatible โ‡„ Claude Messages
  • ๐Ÿ”„ OpenAI Compatible โ†’ Google Gemini
  • ๐Ÿ”„ Google Gemini โ†’ OpenAI Compatible - Text only, function calling not supported yet
  • ๐Ÿšง OpenAI Compatible โ‡„ OpenAI Responses - In development
  • ๐Ÿ”„ Thinking-to-content functionality

Reasoning Effort Support:

View detailed configuration

OpenAI series models:

  • o3-mini-high - High reasoning effort
  • o3-mini-medium - Medium reasoning effort
  • o3-mini-low - Low reasoning effort
  • gpt-5-high - High reasoning effort
  • gpt-5-medium - Medium reasoning effort
  • gpt-5-low - Low reasoning effort

Claude thinking models:

  • claude-3-7-sonnet-20250219-thinking - Enable thinking mode

Google Gemini series models:

  • gemini-2.5-flash-thinking - Enable thinking mode
  • gemini-2.5-flash-nothinking - Disable thinking mode
  • gemini-2.5-pro-thinking - Enable thinking mode
  • gemini-2.5-pro-thinking-128 - Enable thinking mode with thinking budget of 128 tokens
  • You can also append -low, -medium, or -high to any Gemini model name to request the corresponding reasoning effort (no extra thinking-budget suffix needed).

๐Ÿค– Model Support

For details, please refer to API Documentation - Relay Interface

Model Type Description Documentation
๐Ÿค– OpenAI-Compatible OpenAI compatible models Documentation
๐Ÿค– OpenAI Responses OpenAI Responses format Documentation
๐ŸŽจ Midjourney-Proxy Midjourney-Proxy(Plus) Documentation
๐ŸŽต Suno-API Suno API Documentation
๐Ÿ”„ Rerank Cohere, Jina Documentation
๐Ÿ’ฌ Claude Messages format Documentation
๐ŸŒ Gemini Google Gemini format Documentation
๐Ÿ”ง Dify ChatFlow mode -
๐ŸŽฏ Custom Supports complete call address -

๐Ÿ“ก Supported Interfaces

View complete interface list

๐Ÿšข Deployment

Tip

Latest Docker image: calciumion/new-api:latest

๐Ÿ“‹ Deployment Requirements

Component Requirement
Local database SQLite (Docker must mount /data directory)
Remote database MySQL โ‰ฅ 5.7.8 or PostgreSQL โ‰ฅ 9.6
Container engine Docker / Docker Compose

โš™๏ธ Environment Variable Configuration

Common environment variable configuration
Variable Name Description Default Value
SESSION_SECRET Session secret (required for multi-machine deployment) -
CRYPTO_SECRET Encryption secret (required for Redis) -
SQL_DSN Database connection string -
REDIS_CONN_STRING Redis connection string -
STREAMING_TIMEOUT Streaming timeout (seconds) 300
STREAM_SCANNER_MAX_BUFFER_MB Max per-line buffer (MB) for the stream scanner; increase when upstream sends huge image/base64 payloads 64
MAX_REQUEST_BODY_MB Max request body size (MB, counted after decompression; prevents huge requests/zip bombs from exhausting memory). Exceeding it returns 413 32
AZURE_DEFAULT_API_VERSION Azure API version 2025-04-01-preview
ERROR_LOG_ENABLED Error log switch false
PYROSCOPE_URL Pyroscope server address -
PYROSCOPE_APP_NAME Pyroscope application name new-api
PYROSCOPE_BASIC_AUTH_USER Pyroscope basic auth user -
PYROSCOPE_BASIC_AUTH_PASSWORD Pyroscope basic auth password -
PYROSCOPE_MUTEX_RATE Pyroscope mutex sampling rate 5
PYROSCOPE_BLOCK_RATE Pyroscope block sampling rate 5
HOSTNAME Hostname tag for Pyroscope new-api

๐Ÿ“– Complete configuration: Environment Variables Documentation

๐Ÿ”ง Deployment Methods

Method 1: Docker Compose (Recommended)
# Clone the project
git clone https://github.com/QuantumNous/new-api.git
cd new-api

# Edit configuration
nano docker-compose.yml

# Start service
docker-compose up -d
Method 2: Docker Commands

Using SQLite:

docker run --name new-api -d --restart always \
  -p 3000:3000 \
  -e TZ=Asia/Shanghai \
  -v ./data:/data \
  calciumion/new-api:latest

Using MySQL:

docker run --name new-api -d --restart always \
  -p 3000:3000 \
  -e SQL_DSN="root:123456@tcp(localhost:3306)/oneapi" \
  -e TZ=Asia/Shanghai \
  -v ./data:/data \
  calciumion/new-api:latest

๐Ÿ’ก Path explanation:

  • ./data:/data - Relative path, data saved in the data folder of the current directory
  • You can also use absolute path, e.g.: /your/custom/path:/data
Method 3: BaoTa Panel
  1. Install BaoTa Panel (โ‰ฅ 9.2.0 version)
  2. Search for New-API in the application store
  3. One-click installation

๐Ÿ“– Tutorial with images

โš ๏ธ Multi-machine Deployment Considerations

Warning

  • Must set SESSION_SECRET - Otherwise login status inconsistent
  • Shared Redis must set CRYPTO_SECRET - Otherwise data cannot be decrypted

๐Ÿ”„ Channel Retry and Cache

Retry configuration: Settings โ†’ Operation Settings โ†’ General Settings โ†’ Failure Retry Count

Cache configuration:

  • REDIS_CONN_STRING: Redis cache (recommended)
  • MEMORY_CACHE_ENABLED: Memory cache

๐Ÿ”— Related Projects

Upstream Projects

Project Description
One API Original project base
Midjourney-Proxy Midjourney interface support

Supporting Tools

Project Description
neko-api-key-tool Key quota query tool
new-api-horizon New API high-performance optimized version

๐Ÿ’ฌ Help Support

๐Ÿ“– Documentation Resources

Resource Link
๐Ÿ“˜ FAQ FAQ
๐Ÿ’ฌ Community Interaction Communication Channels
๐Ÿ› Issue Feedback Issue Feedback
๐Ÿ“š Complete Documentation Official Documentation

๐Ÿค Contribution Guide

Welcome all forms of contribution!

  • ๐Ÿ› Report Bugs
  • ๐Ÿ’ก Propose New Features
  • ๐Ÿ“ Improve Documentation
  • ๐Ÿ”ง Submit Code

๐Ÿ“œ License

This project is licensed under the GNU Affero General Public License v3.0 (AGPLv3).

This is an open-source project developed based on One API (MIT License).

If your organization's policies do not permit the use of AGPLv3-licensed software, or if you wish to avoid the open-source obligations of AGPLv3, please contact us at: support@quantumnous.com


๐ŸŒŸ Star History

Star History Chart


๐Ÿ’– Thank you for using New API

If this project is helpful to you, welcome to give us a โญ๏ธ Star๏ผ

Official Documentation โ€ข Issue Feedback โ€ข Latest Release

Built with โค๏ธ by QuantumNous

Release History

VersionChangesUrgencyDate
v0.12.14## Improvements - Improved user-visible management and top-up logs with a cleaner display, preserved row expansion, and clearer handling of legacy records. --- **Full Changelog**: https://github.com/QuantumNous/new-api/compare/v0.12.13...v0.12.14High4/17/2026
v0.12.10## New Features - Added passthrough support for Claude `cache_control` and `speed` options, giving you more control over request behavior when using Claude models (#4247). ## Bug Fixes - Fixed OpenAI Responses API compatibility issues when `instructions` contain structured content instead of plain text (#4260). ## Improvements - Improved Stripe payment processing to better handle asynchronous webhook events, making delayed payment confirmations more reliable. - Quota adjustment logs now recordHigh4/15/2026
v0.12.9## New Features - Subscription cards now show the next quota reset time, making it easier to see when usage limits will refresh (#4181). ## Bug Fixes - Claude requests are now more reliable by avoiding incompatible Top P settings when needed. - Azure channels now correctly support the `/v1/responses/compact` route (#4149). - Editing values in GroupTable no longer forces the cursor to jump to the end on each keystroke (#4208). ## Improvements - Dashboard charts have been refined to present rankHigh4/13/2026
nightly-20260409### Highlights / ้‡็‚นๆ›ดๆ–ฐ This release introduces comprehensive support for **tiered billing (้˜ถๆขฏ่ฎก่ดน)**. It allows for flexible pricing expressions, detailed tool call quota calculations, and channel test support for tiered models. ๆœฌๆฌกๅ‘ๅธƒ้‡็‚นๅขžๅŠ ไบ†ๅฏน**้˜ถๆขฏ่ฎก่ดน**็š„ๅ…จ้ขๆ”ฏๆŒ๏ผŒๆ”ฏๆŒ็ตๆดป็š„่ฎก่ดน่กจ่พพๅผ่งฃๆžใ€่‡ชๅฎšไน‰ๅทฅๅ…ท่ฐƒ็”จ้ขๅบฆ่ฎก็ฎ—๏ผŒๅนถๅฎŒๅ–„ไบ†UI้…็ฝฎ้กนๅ’Œๆธ ้“ๆต‹่ฏ•ๆ”ฏๆŒใ€‚ **Docker image:** `calciumion/new-api:nightly` **Docker ้•œๅƒ:** `calciumion/new-api:nightly` <img width="500" alt="" src="https://github.com/user-attachments/assets/893d5789-64b3-414d-8e71-cc178b72High4/9/2026
v0.12.5### New Features * Add support for Minimax image generation relay (#4103). ### Improvements * Redesign group ratio rules with a collapsible layout for better usability. * Enhance dashboard chart axes and sorting logic. * Add a convenient copy button next to the API link in the dashboard info panel. ### Bug Fixes * Resolve email delivery issues for Outlook and similar providers (#4112). <img width="2848" height="966" alt="" src="https://github.com/user-attachments/assets/8cfdd943-3bHigh4/8/2026

Dependencies & License Audit

Loading dependencies...

Similar Packages

llmioLLM API load-balancing gateway. LLM API ่ดŸ่ฝฝๅ‡่กก็ฝ‘ๅ…ณ.v0.8.3
cc-relayโšก๏ธ Blazing fast LLMs API Gateway written in Gov0.0.16
copilot-api๐Ÿš€ Access the reverse-engineered GitHub Copilot API through this proxy, enabling streamlined integration for your development needs.main@2026-04-21
gproxygproxy is a Rust-based multi-channel LLM proxy that exposes OpenAI / Claude / Gemini-style APIs through a unified gateway, with a built-in admin console, user/key management, and request/usage auditinv1.0.18
ai-guideProvide free, open access to comprehensive AI tools, guides, reviews, and resources to reduce knowledge gaps and empower users.main@2026-04-21