Feature Description
- đ¨ Brand new UI interface (some interfaces are still pending updates)
- đ Multi-language support (to be improved)
- đ¨ Added Midjourney-Proxy(Plus) API support
- đ° Supports online top-up functionality, configurable in System Settings:
-
- Epay
-
- đ Supports querying usage Quota by key:
- In conjunction with the project neko-api-key-tool, querying usage by key can be achieved.
- đ Pagination supports selecting the number of items displayed per page
- đ SQLite database storage support, ready to use out of the box, lightweight and convenient
- đĩ Supports model billing by usage count, configurable in System Settings - Operation Settings
- âī¸ Supports weighted random Channel selection
- đ Data Dashboard (Console)
- đ Configurable models that a Token can call
- đ¤ Supports Telegram authorized login:
- System Settings - Configure Login & Registration - Allow login via Telegram
- Enter command /setdomain to @Botfather
- Select your bot, then enter http(s)://your_website_address/login
- The Telegram Bot Name is the bot username string without the @
- đĩ Added Suno API API support
- đ Supports Rerank models, currently compatible with Cohere and Jina, and can be integrated with Dify
- ⥠OpenAI Realtime API - Supports OpenAI's Realtime API, supports Azure Channel
- Supports using the route /chat2link to enter the chat interface
- đ§ Supports setting reasoning effort via model name suffix:
- OpenAI o-series models
-
- Add suffix
-highto set as high reasoning effort (e.g.:o3-mini-high)
- Add suffix
-
- Add suffix
-mediumto set as medium reasoning effort (e.g.:o3-mini-medium)
- Add suffix
-
- Add suffix
-lowto set as low reasoning effort (e.g.:o3-mini-low)
- Add suffix
-
- Claude thinking models
-
- Add suffix
-thinkingto enable thinking mode (e.g.:claude-3-7-sonnet-20250219-thinking)
- Add suffix
-
- OpenAI o-series models
- đ Thinking to Content, supports setting the
thinking_to_contentoption inChannel - Edit - Channel Extra Settings, defaultfalse. When enabled, it will convert the thinking contentreasoning_contentinto a<think>tag and append it to the returned content. - đ Model Rate Limiting, supports setting model rate limits in
System Settings - Rate Limit Settings, including total request count limit and successful request count limit. - đ° Cache Billing Support, when enabled, billing can occur at a set Ratio upon cache hit:
- Set the Prompt Cache Ratio option in
System Settings - Operation Settings - Set the Prompt Cache Ratio in the Channel, range 0-1. For example, setting it to 0.5 means billing at 50% upon cache hit.
- Supported Channels:
-
- OpenAI
-
- Azure
-
- DeepSeek
-
- Claude
-
- Set the Prompt Cache Ratio option in
How is this guide?