Compare commits

..

8 Commits

Author SHA1 Message Date
Vinlic科技
bbe58b70fe
Merge pull request #54 from huliang2016/master
support depth search
2025-01-14 16:15:31 +08:00
胡亮
db50132167 support depth search 2025-01-14 08:06:56 +00:00
Vinlic
9f2e17210a 修复文档和图像上传 2024-12-14 01:10:51 +08:00
Vinlic科技
33b16a3d8c
Merge pull request #52 from Alex-Yanggg/master
Update the Doc and Create the English version
2024-12-12 12:53:18 +08:00
Alex
5e2cb0a8c9
Update README.md 2024-12-12 15:45:12 +11:00
Alex
30153245c7
Create README_EN.md 2024-12-12 15:45:09 +11:00
Alex
2dd2517dc9
Update README.md
添加英文入口与目录层级对齐
2024-12-12 15:06:38 +11:00
Vinlic科技
e95a7e49c6
Update README.md 2024-12-04 17:05:07 +08:00
5 changed files with 565 additions and 37 deletions

View File

@ -1,15 +1,19 @@
# Qwen AI Free 服务 # Qwen AI Free 服务
<hr>
<span>[ 中文 | <a href="README_EN.md">English</a> ]</span>
[![](https://img.shields.io/github/license/llm-red-team/qwen-free-api.svg)](LICENSE) [![](https://img.shields.io/github/license/llm-red-team/qwen-free-api.svg)](LICENSE)
![](https://img.shields.io/github/stars/llm-red-team/qwen-free-api.svg) ![](https://img.shields.io/github/stars/llm-red-team/qwen-free-api.svg)
![](https://img.shields.io/github/forks/llm-red-team/qwen-free-api.svg) ![](https://img.shields.io/github/forks/llm-red-team/qwen-free-api.svg)
![](https://img.shields.io/docker/pulls/vinlic/qwen-free-api.svg) ![](https://img.shields.io/docker/pulls/vinlic/qwen-free-api.svg)
支持高速流式输出、支持多轮对话、支持无水印AI绘图、支持长文档解读、图像解析零配置部署多路token支持自动清理会话痕迹。 支持高速流式输出、支持多轮对话、支持无水印AI绘图、支持长文档解读、图像解析、联网检索零配置部署多路token支持自动清理会话痕迹。
与ChatGPT接口完全兼容。 与ChatGPT接口完全兼容。
还有以下个free-api欢迎关注 还有以下个free-api欢迎关注
Moonshot AIKimi.ai接口转API [kimi-free-api](https://github.com/LLM-Red-Team/kimi-free-api) Moonshot AIKimi.ai接口转API [kimi-free-api](https://github.com/LLM-Red-Team/kimi-free-api)
@ -19,6 +23,10 @@ Moonshot AIKimi.ai接口转API [kimi-free-api](https://github.com/LLM-Red-
秘塔AI (Metaso) 接口转API [metaso-free-api](https://github.com/LLM-Red-Team/metaso-free-api) 秘塔AI (Metaso) 接口转API [metaso-free-api](https://github.com/LLM-Red-Team/metaso-free-api)
字节跳动豆包接口转API [doubao-free-api](https://github.com/LLM-Red-Team/doubao-free-api)
字节跳动即梦AI接口转API [jimeng-free-api](https://github.com/LLM-Red-Team/jimeng-free-api)
讯飞星火Spark接口转API [spark-free-api](https://github.com/LLM-Red-Team/spark-free-api) 讯飞星火Spark接口转API [spark-free-api](https://github.com/LLM-Red-Team/spark-free-api)
MiniMax海螺AI接口转API [hailuo-free-api](https://github.com/LLM-Red-Team/hailuo-free-api) MiniMax海螺AI接口转API [hailuo-free-api](https://github.com/LLM-Red-Team/hailuo-free-api)
@ -29,26 +37,36 @@ MiniMax海螺AI接口转API [hailuo-free-api](https://github.com/LLM-Red-T
## 目录 ## 目录
* [免责声明](#免责声明) - [Qwen AI Free 服务](#qwen-ai-free-服务)
* [在线体验](#在线体验) - [目录](#目录)
* [效果示例](#效果示例) - [免责声明](#免责声明)
* [接入准备](#接入准备) - [效果示例](#效果示例)
* [Docker部署](#Docker部署) - [验明正身Demo](#验明正身demo)
* [Docker-compose部署](#Docker-compose部署) - [多轮对话Demo](#多轮对话demo)
* [Render部署](#Render部署) - [AI绘图Demo](#ai绘图demo)
* [Vercel部署](#Vercel部署) - [长文档解读Demo](#长文档解读demo)
* [原生部署](#原生部署) - [图像解析Demo](#图像解析demo)
* [推荐使用客户端](#推荐使用客户端) - [10线程并发测试](#10线程并发测试)
* [接口列表](#接口列表) - [接入准备](#接入准备)
* [对话补全](#对话补全) - [方法1](#方法1)
* [AI绘图](#AI绘图) - [方法2](#方法2)
* [文档解读](#文档解读) - [多账号接入](#多账号接入)
* [图像解析](#图像解析) - [Docker部署](#docker部署)
* [ticket存活检测](#ticket存活检测) - [Docker-compose部署](#docker-compose部署)
* [注意事项](#注意事项) - [Render部署](#render部署)
* [Nginx反代优化](#Nginx反代优化) - [Vercel部署](#vercel部署)
* [Token统计](#Token统计) - [原生部署](#原生部署)
* [Star History](#star-history) - [推荐使用客户端](#推荐使用客户端)
- [接口列表](#接口列表)
- [对话补全](#对话补全)
- [AI绘图](#ai绘图)
- [文档解读](#文档解读)
- [图像解析](#图像解析)
- [ticket存活检测](#ticket存活检测)
- [注意事项](#注意事项)
- [Nginx反代优化](#nginx反代优化)
- [Token统计](#token统计)
- [Star History](#star-history)
## 免责声明 ## 免责声明
@ -62,12 +80,6 @@ MiniMax海螺AI接口转API [hailuo-free-api](https://github.com/LLM-Red-T
**仅限自用,禁止对外提供服务或商用,避免对官方造成服务压力,否则风险自担!** **仅限自用,禁止对外提供服务或商用,避免对官方造成服务压力,否则风险自担!**
## 在线体验
此链接仅临时测试功能,长期使用请自行部署。
https://udify.app/chat/qOXzVl5kkvhQXM8r
## 效果示例 ## 效果示例
### 验明正身Demo ### 验明正身Demo
@ -84,11 +96,11 @@ https://udify.app/chat/qOXzVl5kkvhQXM8r
### 长文档解读Demo ### 长文档解读Demo
![AI绘图](./doc/example-5.png) ![长文本解读](./doc/example-5.png)
### 图像解析Demo ### 图像解析Demo
![AI绘图](./doc/example-6.png) ![像解析](./doc/example-6.png)
### 10线程并发测试 ### 10线程并发测试
@ -114,9 +126,9 @@ https://udify.app/chat/qOXzVl5kkvhQXM8r
### 多账号接入 ### 多账号接入
你可以通过提供多个账号的tongyi_sso_ticket或login_aliyunid_ticket并使用,拼接提供: 你可以通过提供多个账号的`tongyi_sso_ticket``login_aliyunid_ticket`,并使用,拼接提供:
Authorization: Bearer TOKEN1,TOKEN2,TOKEN3 `Authorization: Bearer TOKEN1,TOKEN2,TOKEN3`
每次请求服务会从中挑选一个。 每次请求服务会从中挑选一个。

510
README_EN.md Normal file
View File

@ -0,0 +1,510 @@
# Qwen AI Free Service
[![](https://img.shields.io/github/license/llm-red-team/qwen-free-api.svg)](LICENSE)
![](https://img.shields.io/github/stars/llm-red-team/qwen-free-api.svg)
![](https://img.shields.io/github/forks/llm-red-team/qwen-free-api.svg)
![](https://img.shields.io/docker/pulls/vinlic/qwen-free-api.svg)
Supports high-speed streaming output, multi-turn dialogues, internet search, long document reading, image analysis, zero-configuration deployment, multi-token support, and automatic session trace cleanup.
Fully compatible with the ChatGPT interface.
Also, the following free APIs are available for your attention:
Moonshot AI (Kimi.ai) API to API [kimi-free-api](https://github.com/LLM-Red-Team/kimi-free-api/tree/master)
StepFun (StepChat) API to API [step-free-api](https://github.com/LLM-Red-Team/step-free-api)
ZhipuAI (ChatGLM) API to API [glm-free-api](https://github.com/LLM-Red-Team/glm-free-api)
Meta Sota (metaso) API to API [metaso-free-api](https://github.com/LLM-Red-Team/metaso-free-api)
Iflytek Spark (Spark) API to API [spark-free-api](https://github.com/LLM-Red-Team/spark-free-api)
Lingxin Intelligence (Emohaa) API to API [emohaa-free-api](https://github.com/LLM-Red-Team/emohaa-free-api) (OUT OF ORDER)
## 目录
* [Announcement](#Announcement)
* [Online experience](#Online-Experience)
* [Effect Examples](#Effect-Examples)
* [Access preparation](#Access-Preparation)
* [Multiple account access](#Multi-Account-Access)
* [Docker Deployment](#Docker-Deployment)
* [Docker-compose Deployment](#Docker-compose-Deployment)
* [Render Deployment](#Render-Deployment)
* [Vercel Deployment](#Vercel-Deployment)
* [Native Deployment](#Native-Deployment)
* [Recommended Clients](#Recommended-Clients)
* [Interface List](#Interface-List)
* [Conversation completion](#conversation-completion)
* [AI Drawing](#AI-Drawing)
* [Document Interpretation](#document-interpretation)
* [Image analysis](#image-analysis)
* [refresh_token survival detection](#refresh_token-survival-detection)
* [Notification](#Notification)
* [Nginx anti-generation optimization](#Nginx-anti-generation-optimization)
* [Token statistics](#Token-statistics)
* [Star History](#star-history)
## Announcement
**This API is unstable. So we highly recommend you go to the [Ali](https://dashscope.console.aliyun.com/) use the offical API, avoiding banned.**
**This organization and individuals do not accept any financial donations and transactions. This project is purely for research, communication, and learning purposes!**
**For personal use only, it is forbidden to provide services or commercial use externally to avoid causing service pressure on the official, otherwise, bear the risk yourself!**
**For personal use only, it is forbidden to provide services or commercial use externally to avoid causing service pressure on the official, otherwise, bear the risk yourself!**
**For personal use only, it is forbidden to provide services or commercial use externally to avoid causing service pressure on the official, otherwise, bear the risk yourself!**
## Online experience
This link is only for temporary testing of functions and cannot be used for a long time. For long-term use, please deploy by yourself.
https://udify.app/chat/qOXzVl5kkvhQXM8r
## Effect Examples
### Identity Verification
![Identity Verification](./doc/example-1.png)
### Multi-turn Dialogue
![Multi-turn Dialogue](./doc/example-2.png)
### AI Drawing
![AI Drawing](./doc/example-3.png)
### Long Document Reading
![Long Document Reading](./doc/example-5.png)
### Image Analysis
![Image Analysis](./doc/example-6.png)
### 10-Thread Concurrency Test
![10-Thread Concurrency Test](./doc/example-4.png)
## Access Preparation
### Method1
Log in to [Tongyi Qianwen](https://tongyi.aliyun.com/qianwen)
Enter Tongyi Qianwen and start a random conversation, then press F12 to open the developer tools. Find the value of `tongyi_sso_ticket` in Application > Cookies, which will be used as the Bearer Token value for Authorization: `Authorization: Bearer TOKEN`
![Get tongyi_sso_ticket](./doc/example-0.png)
### Method2
Log in to [Alibaba Cloud](https://www.aliyun.com/) (not recommended if the account has important assets such as servers). If the account has not previously entered [Tongyi Qianwen](https://tongyi.aliyun.com/qianwen), you need to first agree to the terms, otherwise it will not take effect.
Then press F12 to open the developer tools. Find the value of `login_aliyunid_ticket` in Application > Cookies, which will be used as the Bearer Token value for Authorization: `Authorization: Bearer TOKEN`
![Get login_aliyunid_ticket](./doc/example-7.png)
### Multi-Account Access
You can provide multiple account `tongyi_sso_ticket` or `login_aliyunid_ticket` and use `,` to join them:
`Authorization: Bearer TOKEN1,TOKEN2,TOKEN3`
The service will pick one each time a request is made.
## Docker Deployment
Please prepare a server with a public IP and open port 8000.
Pull the image and start the service
```shell
docker run -it -d --init --name qwen-free-api -p 8000:8000 -e TZ=Asia/Shanghai vinlic/qwen-free-api:latest
```
check real-time service logs
```shell
docker logs -f step-free-api
```
Restart service
```shell
docker restart step-free-api
```
Shut down service
```shell
docker stop step-free-api
```
### Docker-compose Deployment
```yaml
version: '3'
services:
qwen-free-api:
container_name: qwen-free-api
image: vinlic/qwen-free-api:latest
restart: always
ports:
- "8000:8000"
environment:
- TZ=Asia/Shanghai
```
### Render Deployment
**Attention: Some deployment regions may not be able to connect to Kimi. If container logs show request timeouts or connection failures (Singapore has been tested and found unavailable), please switch to another deployment region!**
**Attention: Container instances for free accounts will automatically stop after a period of inactivity, which may result in a 50-second or longer delay during the next request. It is recommended to check [Render Container Keepalive](https://github.com/LLM-Red-Team/free-api-hub/#Render%E5%AE%B9%E5%99%A8%E4%BF%9D%E6%B4%BB)**
1. Fork this project to your GitHub account.
2. Visit [Render](https://dashboard.render.com/) and log in with your GitHub account.
3. Build your Web Service (`New+` -> `Build and deploy from a Git repository` -> `Connect your forked project` -> `Select deployment region` -> `Choose instance type as Free` -> `Create Web Service`).
4. After the build is complete, copy the assigned domain and append the URL to access it.
### Vercel Deployment
**Note: Vercel free accounts have a request response timeout of 10 seconds, but interface responses are usually longer, which may result in a 504 timeout error from Vercel!**
Please ensure that Node.js environment is installed first.
```shell
npm i -g vercel --registry http://registry.npmmirror.com
vercel login
git clone https://github.com/LLM-Red-Team/qwen-free-api
cd qwen-free-api
vercel --prod
```
## Native Deployment
Please prepare a server with a public IP and open port 8000.
Please install the Node.js environment and configure the environment variables first, and confirm that the node command is available.
Install dependencies
```shell
npm i
```
Install PM2 for process guarding
```shell
npm i -g pm2
```
Compile and build. When you see the dist directory, the build is complete.
```shell
npm run build
```
Start service
```shell
pm2 start dist/index.js --name "qwen-free-api"
```
View real-time service logs
```shell
pm2 logs qwen-free-api
```
Restart service
```shell
pm2 reload qwen-free-api
```
Shut down service
```shell
pm2 stop qwen-free-api
```
## Recommended Clients
Using the following second-developed clients for free-api series projects is faster and easier, and supports document/image uploads!
[Clivia](https://github.com/Yanyutin753/lobe-chat)'s modified LobeChat [https://github.com/Yanyutin753/lobe-chat](https://github.com/Yanyutin753/lobe-chat)
[Time@](https://github.com/SuYxh)'s modified ChatGPT Web [https://github.com/SuYxh/chatgpt-web-sea](https://github.com/SuYxh/chatgpt-web-sea)
## interface list
Currently, the `/v1/chat/completions` interface compatible with openai is supported. You can use the client access interface compatible with openai or other clients, or use online services such as [dify](https://dify.ai/) Access and use.
### Conversation completion
Conversation completion interface, compatible with openai's [chat-completions-api](https://platform.openai.com/docs/guides/text-generation/chat-completions-api).
**POST /v1/chat/completions**
The header needs to set the Authorization header:
```
Authorization: Bearer [tongyi_sso_ticket/login_aliyunid_ticket]
```
Request data:
```json
{
// Fill in the model name as you like.
"model": "qwen",
// Currently, multi-round conversations are realized based on message merging, which in some scenarios may lead to capacity degradation and is limited by the maximum number of tokens in a single round.
// If you want a native multi-round dialog experience, you can pass in the ids obtained from the last round of messages to pick up the context
// "conversation_id": "bc9ef150d0e44794ab624df958292300-40811965812e4782bb87f1a9e4e2b2cd",
"messages": [
{
"role": "user",
"content": "who RU"
}
],
// If using SSE stream, please set it to true, the default is false
"stream": false
}
```
Response data:
```json
{
// For a native multi-round conversation experience, this id, you can pass in the conversation_id for the next round of conversation to pick up the context
"id": "bc9ef150d0e44794ab624df958292300-40811965812e4782bb87f1a9e4e2b2cd",
"model": "qwen",
"object": "chat.completion",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "I'm Qwen."
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 1,
"completion_tokens": 1,
"total_tokens": 2
},
"created": 1710152062
}
```
### AI Drawing
Conversation completion interface, compatible with openai's [chat-completions-api](https://platform.openai.com/docs/guides/text-generation/chat-completions-api).
**POST /v1/chat/completions**
The header needs to set the Authorization header:
```
Authorization: Bearer [tongyi_sso_ticket/login_aliyunid_ticket]
```
Request data
```json
{
// 可以乱填
"model": "wanxiang",
"prompt": "A cut cat"
}
```
Response data
```json
{
"created": 1711507734,
"data": [
{
"url": "https://wanx.alicdn.com/wanx/1111111111/text_to_image/7248e85cfda6491aae59c54e7e679b17_0.png"
}
]
}
```
### Document interpretation
Provide an accessible file URL or BASE64_URL to parse.
**POST /v1/chat/completions**
The header needs to set the Authorization header:
```
Authorization: Bearer [refresh_token]
```
Request data:
```json
{
"model": "qwen",
"messages": [
{
"role": "user",
"content": [
{
"type": "file",
"file_url": {
"url": "https://mj101-1317487292.cos.ap-shanghai.myqcloud.com/ai/test.pdf"
}
},
{
"type": "text",
"text": "What does the document say"
}
]
}
]
}
```
Response data:
```json
{
"id": "b56ea6c9e86140429fa2de6a6ec028ff",
"model": "qwen",
"object": "chat.completion",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "This is a doc about the magic of love. balabala..."
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 1,
"completion_tokens": 1,
"total_tokens": 2
},
"created": 1712253736
}
```
### Image analysis
Provide an accessible image URL or BASE64_URL to parse.
This format is compatible with the [gpt-4-vision-preview](https://platform.openai.com/docs/guides/vision) API format. You can also use this format to transmit documents for parsing.
**POST /v1/chat/completions**
The header needs to set the Authorization header:
```
Authorization: Bearer [refresh_token]
```
请求数据:
```json
{
"model": "qwen",
"messages": [
{
"role": "user",
"content": [
{
"type": "file",
"file_url": {
"url": "https://img.alicdn.com/imgextra/i1/O1CN01CC9kic1ig1r4sAY5d_!!6000000004441-2-tps-880-210.png"
}
},
{
"type": "text",
"text": "What does the image describe?"
}
]
}
]
}
```
Response data:
```json
{
"id": "895fbe7fa22442d499ba67bb5213e842",
"model": "qwen",
"object": "chat.completion",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "It's the logo of Qwen."
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 1,
"completion_tokens": 1,
"total_tokens": 2
},
"created": 1712254066
}
```
### refresh_token survival detection
Check whether refresh_token is alive. If live is not true, otherwise it is false. Please do not call this interface frequently (less than 10 minutes).
**POST /token/check**
Request data:
```json
{
"token": "QIhaHrrXUaIrWMUmL..."
}
```
Response data:
```json
{
"live": true
}
```
## Notification
### Nginx anti-generation optimization
If you are using Nginx reverse proxy `qwen-free-api`, please add the following configuration items to optimize the output effect of the stream and optimize the experience.
```nginx
# Turn off proxy buffering. When set to off, Nginx will immediately send client requests to the backend server and immediately send responses received from the backend server back to the client.
proxy_buffering off;
# Enable chunked transfer encoding. Chunked transfer encoding allows servers to send data in chunks for dynamically generated content without knowing the size of the content in advance.
chunked_transfer_encoding on;
# Turn on TCP_NOPUSH, which tells Nginx to send as much data as possible before sending the packet to the client. This is usually used in conjunction with sendfile to improve network efficiency.
tcp_nopush on;
# Turn on TCP_NODELAY, which tells Nginx not to delay sending data and to send small data packets immediately. In some cases, this can reduce network latency.
tcp_nodelay on;
#Set the timeout to keep the connection, here it is set to 120 seconds. If there is no further communication between client and server during this time, the connection will be closed.
keepalive_timeout 120;
```
### Token statistics
Since the inference side is not in qwen-free-api, the token cannot be counted and will be returned as a fixed number!!!!!
## Star History
[![Star History Chart](https://api.star-history.com/svg?repos=LLM-Red-Team/qwen-free-api&type=Date)](https://star-history.com/#LLM-Red-Team/qwen-free-api&Date)

View File

@ -1,6 +1,6 @@
{ {
"name": "qwen-free-api", "name": "qwen-free-api",
"version": "0.0.20", "version": "0.0.21",
"description": "Qwen Free API Server", "description": "Qwen Free API Server",
"type": "module", "type": "module",
"main": "dist/index.js", "main": "dist/index.js",

View File

@ -80,6 +80,7 @@ async function removeConversation(convId: string, ticket: string) {
async function createCompletion( async function createCompletion(
model = MODEL_NAME, model = MODEL_NAME,
messages: any[], messages: any[],
searchType: string = '',
ticket: string, ticket: string,
refConvId = '', refConvId = '',
retryCount = 0 retryCount = 0
@ -129,7 +130,8 @@ async function createCompletion(
sessionType: "text_chat", sessionType: "text_chat",
parentMsgId, parentMsgId,
params: { params: {
"fileUploadBatchId": util.uuid() "fileUploadBatchId": util.uuid(),
"searchType": searchType,
}, },
contents: messagesPrepare(messages, refs, !!refConvId), contents: messagesPrepare(messages, refs, !!refConvId),
}) })
@ -173,6 +175,7 @@ async function createCompletion(
async function createCompletionStream( async function createCompletionStream(
model = MODEL_NAME, model = MODEL_NAME,
messages: any[], messages: any[],
searchType: string = '',
ticket: string, ticket: string,
refConvId = '', refConvId = '',
retryCount = 0 retryCount = 0
@ -220,7 +223,8 @@ async function createCompletionStream(
sessionType: "text_chat", sessionType: "text_chat",
parentMsgId, parentMsgId,
params: { params: {
"fileUploadBatchId": util.uuid() "fileUploadBatchId": util.uuid(),
"searchType": searchType,
}, },
contents: messagesPrepare(messages, refs, !!refConvId), contents: messagesPrepare(messages, refs, !!refConvId),
}) })
@ -788,7 +792,7 @@ async function uploadFile(fileUrl: string, ticket: string) {
// 上传文件到OSS // 上传文件到OSS
await axios.request({ await axios.request({
method: "POST", method: "POST",
url: "https://broadscope-dialogue.oss-cn-beijing.aliyuncs.com/", url: "https://broadscope-dialogue-new.oss-cn-beijing.aliyuncs.com/",
data: formData, data: formData,
// 100M限制 // 100M限制
maxBodyLength: FILE_MAX_SIZE, maxBodyLength: FILE_MAX_SIZE,

View File

@ -17,11 +17,12 @@ export default {
const tokens = chat.tokenSplit(request.headers.authorization); const tokens = chat.tokenSplit(request.headers.authorization);
// 随机挑选一个ticket // 随机挑选一个ticket
const token = _.sample(tokens); const token = _.sample(tokens);
const { model, conversation_id: convId, messages, stream } = request.body; const { model, conversation_id: convId, messages, search_type, stream } = request.body;
if (stream) { if (stream) {
const stream = await chat.createCompletionStream( const stream = await chat.createCompletionStream(
model, model,
messages, messages,
search_type,
token, token,
convId convId
); );
@ -32,6 +33,7 @@ export default {
return await chat.createCompletion( return await chat.createCompletion(
model, model,
messages, messages,
search_type,
token, token,
convId convId
); );