Compare commits

..

No commits in common. "master" and "0.0.31" have entirely different histories.

9 changed files with 285 additions and 2126 deletions

View File

@ -11,11 +11,11 @@
![](https://img.shields.io/github/forks/llm-red-team/kimi-free-api.svg) ![](https://img.shields.io/github/forks/llm-red-team/kimi-free-api.svg)
![](https://img.shields.io/docker/pulls/vinlic/kimi-free-api.svg) ![](https://img.shields.io/docker/pulls/vinlic/kimi-free-api.svg)
支持高速流式输出、支持多轮对话、支持联网搜索、支持智能体对话、支持探索版、支持K1思考模型、支持长文档解读、支持图像解析零配置部署多路token支持自动清理会话痕迹。 支持高速流式输出、支持多轮对话、支持联网搜索、支持长文档解读、支持图像解析零配置部署多路token支持自动清理会话痕迹。
与ChatGPT接口完全兼容。 与ChatGPT接口完全兼容。
还有以下个free-api欢迎关注 还有以下个free-api欢迎关注
阶跃星辰 (跃问StepChat) 接口转API [step-free-api](https://github.com/LLM-Red-Team/step-free-api) 阶跃星辰 (跃问StepChat) 接口转API [step-free-api](https://github.com/LLM-Red-Team/step-free-api)
@ -25,29 +25,22 @@
秘塔AI (Metaso) 接口转API [metaso-free-api](https://github.com/LLM-Red-Team/metaso-free-api) 秘塔AI (Metaso) 接口转API [metaso-free-api](https://github.com/LLM-Red-Team/metaso-free-api)
字节跳动豆包接口转API [doubao-free-api](https://github.com/LLM-Red-Team/doubao-free-api)
字节跳动即梦AI接口转API [jimeng-free-api](https://github.com/LLM-Red-Team/jimeng-free-api)
讯飞星火Spark接口转API [spark-free-api](https://github.com/LLM-Red-Team/spark-free-api) 讯飞星火Spark接口转API [spark-free-api](https://github.com/LLM-Red-Team/spark-free-api)
MiniMax海螺AI接口转API [hailuo-free-api](https://github.com/LLM-Red-Team/hailuo-free-api) 聆心智能 (Emohaa) 接口转API [emohaa-free-api](https://github.com/LLM-Red-Team/emohaa-free-api)
深度求索DeepSeek接口转API [deepseek-free-api](https://github.com/LLM-Red-Team/deepseek-free-api)
聆心智能 (Emohaa) 接口转API [emohaa-free-api](https://github.com/LLM-Red-Team/emohaa-free-api)(当前不可用)
## 目录 ## 目录
* [免责声明](#免责声明) * [免责声明](#免责声明)
* [在线体验](#在线体验)
* [效果示例](#效果示例) * [效果示例](#效果示例)
* [接入准备](#接入准备) * [接入准备](#接入准备)
* [多账号接入](#多账号接入) * [多账号接入](#多账号接入)
* [Docker部署](#Docker部署) * [Docker部署](#Docker部署)
* [Docker-compose部署](#Docker-compose部署) * [Docker-compose部署](#Docker-compose部署)
* [Render部署](#Render部署) * [Render部署](#Render部署)
* [Vercel部署](#Vercel部署) * [Vercel部署](#Vercel部署)
* [Zeabur部署](#Zeabur部署) * [Zeabur部署](#Zeabur部署)
* [原生部署](#原生部署) * [原生部署](#原生部署)
* [推荐使用客户端](#推荐使用客户端) * [推荐使用客户端](#推荐使用客户端)
* [接口列表](#接口列表) * [接口列表](#接口列表)
@ -72,6 +65,12 @@ MiniMax海螺AI接口转API [hailuo-free-api](https://github.com/LLM-Red-T
**仅限自用,禁止对外提供服务或商用,避免对官方造成服务压力,否则风险自担!** **仅限自用,禁止对外提供服务或商用,避免对官方造成服务压力,否则风险自担!**
## 在线体验
此链接仅临时测试功能,不可长期使用,长期使用请自行部署。
https://udify.app/chat/Po0F6BMJ15q5vu2P
## 效果示例 ## 效果示例
### 验明正身Demo ### 验明正身Demo
@ -86,17 +85,11 @@ MiniMax海螺AI接口转API [hailuo-free-api](https://github.com/LLM-Red-T
![联网搜索](./doc/example-2.png) ![联网搜索](./doc/example-2.png)
### 智能体对话Demo
此处使用 [翻译通](https://kimi.moonshot.cn/chat/coo6l3pkqq4ri39f36bg) 智能体。
![智能体对话](./doc/example-7.png)
### 长文档解读Demo ### 长文档解读Demo
![长文档解读](./doc/example-5.png) ![长文档解读](./doc/example-5.png)
### 图像OCR Demo ### 图像解析Demo
![图像解析](./doc/example-3.png) ![图像解析](./doc/example-3.png)
@ -126,7 +119,7 @@ MiniMax海螺AI接口转API [hailuo-free-api](https://github.com/LLM-Red-T
## Docker部署 ## Docker部署
请准备能够部署Docker镜像且能够访问网络的设备或服务器并将8000端口开放。 请准备一台具有公网IP的服务器并将8000端口开放。
拉取镜像并启动服务 拉取镜像并启动服务
@ -276,15 +269,7 @@ Authorization: Bearer [refresh_token]
请求数据: 请求数据:
```json ```json
{ {
// 模型名称 // 模型名称随意填写如果不希望输出检索过程模型名称请包含silent_search
// kimi默认模型
// kimi-search联网检索模型
// kimi-research探索版模型
// kimi-k1K1模型
// kimi-math数学模型
// kimi-silent不输出检索过程模型
// search/research/k1/math/silent可自由组合使用
// 如果使用kimi+智能体model请填写智能体ID就是浏览器地址栏上尾部的一串英文+数字20个字符的ID
"model": "kimi", "model": "kimi",
// 目前多轮对话基于消息合并实现某些场景可能导致能力下降且受单轮最大Token数限制 // 目前多轮对话基于消息合并实现某些场景可能导致能力下降且受单轮最大Token数限制
// 如果您想获得原生的多轮对话体验可以传入首轮消息获得的id来接续上下文注意如果使用这个首轮必须传none否则第二轮会空响应 // 如果您想获得原生的多轮对话体验可以传入首轮消息获得的id来接续上下文注意如果使用这个首轮必须传none否则第二轮会空响应
@ -343,15 +328,7 @@ Authorization: Bearer [refresh_token]
请求数据: 请求数据:
```json ```json
{ {
// 模型名称 // 模型名称随意填写如果不希望输出检索过程模型名称请包含silent_search
// kimi默认模型
// kimi-search联网检索模型
// kimi-research探索版模型
// kimi-k1K1模型
// kimi-math数学模型
// kimi-silent不输出检索过程模型
// search/research/k1/math/silent可自由组合使用
// 如果使用kimi+智能体model请填写智能体ID就是浏览器地址栏上尾部的一串英文+数字20个字符的ID
"model": "kimi", "model": "kimi",
"messages": [ "messages": [
{ {
@ -400,7 +377,7 @@ Authorization: Bearer [refresh_token]
} }
``` ```
### 图像OCR ### 图像解析
提供一个可访问的图像URL或者BASE64_URL进行解析。 提供一个可访问的图像URL或者BASE64_URL进行解析。
@ -417,15 +394,7 @@ Authorization: Bearer [refresh_token]
请求数据: 请求数据:
```json ```json
{ {
// 模型名称 // 模型名称随意填写如果不希望输出检索过程模型名称请包含silent_search
// kimi默认模型
// kimi-search联网检索模型
// kimi-research探索版模型
// kimi-k1K1模型
// kimi-math数学模型
// kimi-silent不输出检索过程模型
// search/research/k1/math/silent可自由组合使用
// 如果使用kimi+智能体model请填写智能体ID就是浏览器地址栏上尾部的一串英文+数字20个字符的ID
"model": "kimi", "model": "kimi",
"messages": [ "messages": [
{ {

View File

@ -12,7 +12,7 @@ Supports high-speed streaming output, multi-turn dialogues, internet search, lon
Fully compatible with the ChatGPT interface. Fully compatible with the ChatGPT interface.
Also, the following free APIs are available for your attention: Also, the following six free APIs are available for your attention:
StepFun (StepChat) API to API [step-free-api](https://github.com/LLM-Red-Team/step-free-api) StepFun (StepChat) API to API [step-free-api](https://github.com/LLM-Red-Team/step-free-api)
@ -24,34 +24,30 @@ Meta Sota (metaso) API to API [metaso-free-api](https://github.com/LLM-Red-Team/
Iflytek Spark (Spark) API to API [spark-free-api](https://github.com/LLM-Red-Team/spark-free-api) Iflytek Spark (Spark) API to API [spark-free-api](https://github.com/LLM-Red-Team/spark-free-api)
Lingxin Intelligence (Emohaa) API to API [emohaa-free-api](https://github.com/LLM-Red-Team/emohaa-free-api) (OUT OF ORDER) Lingxin Intelligence (Emohaa) API to API [emohaa-free-api](https://github.com/LLM-Red-Team/emohaa-free-api)
## Table of Contents ## Table of Contents
* [Announcement](#Announcement) * [Disclaimer](#disclaimer)
* [Online experience](#Online-Experience) * [Online experience](#Online-Experience)
* [Effect Examples](#Effect-Examples) * [Effect Examples](#Effect-Examples)
* [Access preparation](#Access-Preparation) * [Access preparation](#Access-Preparation)
* [Multiple account access](#Multi-Account-Access) * [Multiple account access](#Multi-Account-Access)
* [Docker Deployment](#Docker-Deployment) * [Docker Deployment](#Docker-Deployment)
* [Docker-compose Deployment](#Docker-compose-Deployment) * [Docker-compose deployment](#Docker-compose-deployment)
* [Render Deployment](Render-Deployment) * [Zeabur Deployment](#Zeabur-Deployment)
* [Vercel Deployment](#Vercel-Deployment) * [Native Deployment](#Native-deployment)
* [Zeabur Deployment](#Zeabur-Deployment)
* [Native Deployment](#Native-Deployment)
* [Interface List](#Interface-List) * [Interface List](#Interface-List)
* [Conversation completion](#conversation-completion) * [Conversation completion](#conversation-completion)
* [Document Interpretation](#document-interpretation) * [Document Interpretation](#document-interpretation)
* [Image analysis](#image-analysis) * [Image analysis](#image-analysis)
* [refresh_token survival detection](#refresh_token-survival-detection) * [refresh_token survival detection](#refresh_token-survival-detection)
* [Precautions](#Precautions) * [Precautions](#Precautions)
* [Nginx anti-generation optimization](#Nginx-anti-generation-optimization) * [Nginx anti-generation optimization](#Nginx-anti-generation-optimization)
* [Token statistics](#Token-statistics) * [Token statistics](#Token-statistics)
* [Star History](#star-history) * [Star History](#star-history)
## Announcement ## Disclaimer
**This API is unstable. So we highly recommend you go to the [MoonshotAI](https://platform.moonshot.cn/) use the offical API, avoiding banned.**
**This organization and individuals do not accept any financial donations and transactions. This project is purely for research, communication, and learning purposes!** **This organization and individuals do not accept any financial donations and transactions. This project is purely for research, communication, and learning purposes!**
@ -135,13 +131,13 @@ Restart service
docker restart kimi-free-api docker restart kimi-free-api
``` ```
Shut down service Out of service
```shell ```shell
docker stop kimi-free-api docker stop kimi-free-api
``` ```
### Docker-compose Deployment ### Docker-compose deployment
```yaml ```yaml
version: '3' version: '3'
@ -157,37 +153,11 @@ services:
- TZ=Asia/Shanghai - TZ=Asia/Shanghai
``` ```
### Render Deployment ## Zeabur Deployment
**Attention: Some deployment regions may not be able to connect to Kimi. If container logs show request timeouts or connection failures (Singapore has been tested and found unavailable), please switch to another deployment region!**
**Attention Container instances for free accounts will automatically stop after a period of inactivity, which may result in a 50-second or longer delay during the next request. It is recommended to check [Render Container Keepalive](https://github.com/LLM-Red-Team/free-api-hub/#Render%E5%AE%B9%E5%99%A8%E4%BF%9D%E6%B4%BB)**
1. Fork this project to your GitHub account.
2. Visit [Render](https://dashboard.render.com/) and log in with your GitHub account.
3. Build your Web Service (New+ -> Build and deploy from a Git repository -> Connect your forked project -> Select deployment region -> Choose instance type as Free -> Create Web Service).
4. After the build is complete, copy the assigned domain and append the URL to access it.
### Vercel Deployment
**Note: Vercel free accounts have a request response timeout of 10 seconds, but interface responses are usually longer, which may result in a 504 timeout error from Vercel!**
Please ensure that Node.js environment is installed first.
```shell
npm i -g vercel --registry http://registry.npmmirror.com
vercel login
git clone https://github.com/LLM-Red-Team/kimi-free-api
cd kimi-free-api
vercel --prod
```
### Zeabur Deployment
[![Deploy on Zeabur](https://zeabur.com/button.svg)](https://zeabur.com/templates/GRFYBP) [![Deploy on Zeabur](https://zeabur.com/button.svg)](https://zeabur.com/templates/GRFYBP)
## Native Deployment ## Native deployment
Please prepare a server with a public IP and open port 8000. Please prepare a server with a public IP and open port 8000.
@ -229,20 +199,12 @@ Restart service
pm2 reload kimi-free-api pm2 reload kimi-free-api
``` ```
Shut down service Out of service
```shell ```shell
pm2 stop kimi-free-api pm2 stop kimi-free-api
``` ```
## Recommended Clients
Using the following second-developed clients for free-api series projects is faster and easier, and supports document/image uploads!
[Clivia](https://github.com/Yanyutin753/lobe-chat)'s modified LobeChat [https://github.com/Yanyutin753/lobe-chat](https://github.com/Yanyutin753/lobe-chat)
[Time@](https://github.com/SuYxh)'s modified ChatGPT Web [https://github.com/SuYxh/chatgpt-web-sea](https://github.com/SuYxh/chatgpt-web-sea)
## interface list ## interface list
Currently, the `/v1/chat/completions` interface compatible with openai is supported. You can use the client access interface compatible with openai or other clients, or use online services such as [dify](https://dify.ai/) Access and use. Currently, the `/v1/chat/completions` interface compatible with openai is supported. You can use the client access interface compatible with openai or other clients, or use online services such as [dify](https://dify.ai/) Access and use.
@ -261,15 +223,7 @@ Authorization: Bearer [refresh_token]
Request data: Request data:
```json ```json
{ {
// Model name // Fill in the model name as you like. If you do not want to output the retrieval process model name, please include silent_search.
// kimi: default model
// kimi-search: online search model
// kimi-research: exploration version model
// kimi-k1: K1 model
// kimi-math: math model
// kimi-silent: model without search process output
// search/research/k1/math/silent: can be freely combined
// If using kimi+agent, fill in the agent ID for model, which is the 20-character ID of letters and numbers at the end of the browser address bar
"model": "kimi", "model": "kimi",
"messages": [ "messages": [
{ {
@ -325,35 +279,27 @@ Authorization: Bearer [refresh_token]
Request data: Request data:
```json ```json
{ {
// Model name // Fill in the model name as you like. If you do not want to output the retrieval process model name, please include silent_search.
// kimi: default model "model": "kimi",
// kimi-search: online search model "messages": [
// kimi-research: exploration version model {
// kimi-k1: K1 model "role": "user",
// kimi-math: math model "content": [
// kimi-silent: model without search process output {
// search/research/k1/math/silent: can be freely combined "type": "file",
// If using kimi+agent, fill in the agent ID for model, which is the 20-character ID of letters and numbers at the end of the browser address bar "file_url": {
"model": "kimi", "url": "https://mj101-1317487292.cos.ap-shanghai.myqcloud.com/ai/test.pdf"
"messages": [ }
{ },
"role": "user", {
"content": [ "type": "text",
{ "text": "What does the document say?"
"type": "file", }
"file_url": { ]
"url": "https://mj101-1317487292.cos.ap-shanghai.myqcloud.com/ai/test.pdf" }
} ],
}, // It is recommended to turn off online search to prevent interference in interpreting results.
{ "use_search": false
"type": "text",
"text": "What does the document say?"
}
]
}
],
// It is recommended to turn off online search to prevent interference in interpreting results.
"use_search": false
} }
``` ```
@ -399,15 +345,7 @@ Authorization: Bearer [refresh_token]
Request data: Request data:
```json ```json
{ {
// Model name // Fill in the model name as you like. If you do not want to output the retrieval process model name, please include silent_search.
// kimi: default model
// kimi-search: online search model
// kimi-research: exploration version model
// kimi-k1: K1 model
// kimi-math: math model
// kimi-silent: model without search process output
// search/research/k1/math/silent: can be freely combined
// If using kimi+agent, fill in the agent ID for model, which is the 20-character ID of letters and numbers at the end of the browser address bar
"model": "kimi", "model": "kimi",
"messages": [ "messages": [
{ {
@ -476,7 +414,7 @@ Response data:
} }
``` ```
## Notification ## Precautions
### Nginx anti-generation optimization ### Nginx anti-generation optimization

Binary file not shown.

Before

Width:  |  Height:  |  Size: 44 KiB

View File

@ -1,6 +1,6 @@
{ {
"name": "kimi-free-api", "name": "kimi-free-api",
"version": "0.0.36", "version": "0.0.31",
"description": "Kimi Free API Server", "description": "Kimi Free API Server",
"type": "module", "type": "module",
"main": "dist/index.js", "main": "dist/index.js",
@ -13,8 +13,8 @@
"dist/" "dist/"
], ],
"scripts": { "scripts": {
"dev": "tsup src/index.ts --format cjs,esm --sourcemap --dts --publicDir public --watch --onSuccess \"node --enable-source-maps --no-node-snapshot dist/index.js\"", "dev": "tsup src/index.ts --format cjs,esm --sourcemap --dts --publicDir public --watch --onSuccess \"node dist/index.js\"",
"start": "node --enable-source-maps --no-node-snapshot dist/index.js", "start": "node dist/index.js",
"build": "tsup src/index.ts --format cjs,esm --sourcemap --dts --clean --publicDir public" "build": "tsup src/index.ts --format cjs,esm --sourcemap --dts --clean --publicDir public"
}, },
"author": "Vinlic", "author": "Vinlic",

View File

@ -5,6 +5,5 @@ export default {
API_TOKEN_EXPIRES: [-2002, 'Token已失效'], API_TOKEN_EXPIRES: [-2002, 'Token已失效'],
API_FILE_URL_INVALID: [-2003, '远程文件URL非法'], API_FILE_URL_INVALID: [-2003, '远程文件URL非法'],
API_FILE_EXECEEDS_SIZE: [-2004, '远程文件超出大小'], API_FILE_EXECEEDS_SIZE: [-2004, '远程文件超出大小'],
API_CHAT_STREAM_PUSHING: [-2005, '已有对话流正在输出'], API_CHAT_STREAM_PUSHING: [-2005, '已有对话流正在输出']
API_RESEARCH_EXCEEDS_LIMIT: [-2006, '探索版使用量已达到上限']
} }

View File

@ -2,9 +2,8 @@ import { PassThrough } from "stream";
import path from 'path'; import path from 'path';
import _ from 'lodash'; import _ from 'lodash';
import mime from 'mime'; import mime from 'mime';
import axios, { AxiosRequestConfig, AxiosResponse } from 'axios'; import axios, { AxiosResponse } from 'axios';
import type IStreamMessage from "../interfaces/IStreamMessage.ts";
import APIException from "@/lib/exceptions/APIException.ts"; import APIException from "@/lib/exceptions/APIException.ts";
import EX from "@/api/consts/exceptions.ts"; import EX from "@/api/consts/exceptions.ts";
import { createParser } from 'eventsource-parser' import { createParser } from 'eventsource-parser'
@ -13,39 +12,27 @@ import util from '@/lib/util.ts';
// 模型名称 // 模型名称
const MODEL_NAME = 'kimi'; const MODEL_NAME = 'kimi';
// 设备ID
const DEVICE_ID = Math.random() * 999999999999999999 + 7000000000000000000;
// SessionID
const SESSION_ID = Math.random() * 99999999999999999 + 1700000000000000000;
// access_token有效期 // access_token有效期
const ACCESS_TOKEN_EXPIRES = 300; const ACCESS_TOKEN_EXPIRES = 300;
// 最大重试次数 // 最大重试次数
const MAX_RETRY_COUNT = 3; const MAX_RETRY_COUNT = 3;
// 重试延迟 // 重试延迟
const RETRY_DELAY = 5000; const RETRY_DELAY = 5000;
// 基础URL
const BASE_URL = 'https://kimi.moonshot.cn';
// 伪装headers // 伪装headers
const FAKE_HEADERS = { const FAKE_HEADERS = {
'Accept': '*/*', 'Accept': '*/*',
'Accept-Encoding': 'gzip, deflate, br, zstd', 'Accept-Encoding': 'gzip, deflate, br, zstd',
'Accept-Language': 'zh-CN,zh;q=0.9,en-US;q=0.8,en;q=0.7', 'Accept-Language': 'zh-CN,zh;q=0.9',
'Cache-Control': 'no-cache', 'Origin': 'https://kimi.moonshot.cn',
'Pragma': 'no-cache',
'Origin': BASE_URL,
'Cookie': util.generateCookie(), 'Cookie': util.generateCookie(),
'R-Timezone': 'Asia/Shanghai', 'R-Timezone': 'Asia/Shanghai',
'Sec-Ch-Ua': '"Google Chrome";v="131", "Chromium";v="131", "Not_A Brand";v="24"', 'Sec-Ch-Ua': '"Google Chrome";v="123", "Not:A-Brand";v="8", "Chromium";v="123"',
'Sec-Ch-Ua-Mobile': '?0', 'Sec-Ch-Ua-Mobile': '?0',
'Sec-Ch-Ua-Platform': '"Windows"', 'Sec-Ch-Ua-Platform': '"Windows"',
'Sec-Fetch-Dest': 'empty', 'Sec-Fetch-Dest': 'empty',
'Sec-Fetch-Mode': 'cors', 'Sec-Fetch-Mode': 'cors',
'Sec-Fetch-Site': 'same-origin', 'Sec-Fetch-Site': 'same-origin',
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/131.0.0.0 Safari/537.36', 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/123.0.0.0 Safari/537.36'
'Priority': 'u=1, i',
'X-Msh-Device-Id': `${DEVICE_ID}`,
'X-Msh-Platform': 'web',
'X-Msh-Session-Id': `${SESSION_ID}`
}; };
// 文件最大大小 // 文件最大大小
const FILE_MAX_SIZE = 100 * 1024 * 1024; const FILE_MAX_SIZE = 100 * 1024 * 1024;
@ -67,10 +54,23 @@ async function requestToken(refreshToken: string) {
accessTokenRequestQueueMap[refreshToken] = []; accessTokenRequestQueueMap[refreshToken] = [];
logger.info(`Refresh token: ${refreshToken}`); logger.info(`Refresh token: ${refreshToken}`);
const result = await (async () => { const result = await (async () => {
const result = await axios.get(`${BASE_URL}/api/auth/token/refresh`, { const result = await axios.get('https://kimi.moonshot.cn/api/auth/token/refresh', {
headers: { headers: {
Accept: '*/*',
'Accept-Encoding': 'gzip, deflate, br, zstd',
'Accept-Language': 'zh-CN,zh;q=0.9',
Authorization: `Bearer ${refreshToken}`, Authorization: `Bearer ${refreshToken}`,
...FAKE_HEADERS, 'Cache-Control': 'no-cache',
'Cookie': util.generateCookie(),
Pragma: 'no-cache',
Referer: 'https://kimi.moonshot.cn/',
'Sec-Ch-Ua': '"Google Chrome";v="123", "Not:A-Brand";v="8", "Chromium";v="123"',
'Sec-Ch-Ua-Mobile': '?0',
'Sec-Ch-Ua-Platform': '"Windows"',
'Sec-Fetch-Dest': 'empty',
'Sec-Fetch-Mode': 'cors',
'Sec-Fetch-Site': 'same-origin',
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/122.0.0.0 Safari/537.36'
}, },
timeout: 15000, timeout: 15000,
validateStatus: () => true validateStatus: () => true
@ -79,18 +79,9 @@ async function requestToken(refreshToken: string) {
access_token, access_token,
refresh_token refresh_token
} = checkResult(result, refreshToken); } = checkResult(result, refreshToken);
const userResult = await axios.get(`${BASE_URL}/api/user`, { const { id: userId } = await getUserInfo(access_token, refreshToken);
headers: {
Authorization: `Bearer ${access_token}`,
...FAKE_HEADERS,
},
timeout: 15000,
validateStatus: () => true
});
if(!userResult.data.id)
throw new APIException(EX.API_REQUEST_FAILED, '获取用户信息失败');
return { return {
userId: userResult.data.id, userId,
accessToken: access_token, accessToken: access_token,
refreshToken: refresh_token, refreshToken: refresh_token,
refreshTime: util.unixTimestamp() + ACCESS_TOKEN_EXPIRES refreshTime: util.unixTimestamp() + ACCESS_TOKEN_EXPIRES
@ -105,7 +96,6 @@ async function requestToken(refreshToken: string) {
return result; return result;
}) })
.catch(err => { .catch(err => {
logger.error(err);
if (accessTokenRequestQueueMap[refreshToken]) { if (accessTokenRequestQueueMap[refreshToken]) {
accessTokenRequestQueueMap[refreshToken].forEach(resolve => resolve(err)); accessTokenRequestQueueMap[refreshToken].forEach(resolve => resolve(err));
delete accessTokenRequestQueueMap[refreshToken]; delete accessTokenRequestQueueMap[refreshToken];
@ -138,32 +128,19 @@ async function acquireToken(refreshToken: string): Promise<any> {
} }
/** /**
* *
*
* @param refreshToken access_token的refresh_token
*/ */
export async function request( async function getUserInfo(accessToken: string, refreshToken: string) {
method: string, const result = await axios.get('https://kimi.moonshot.cn/api/user', {
uri: string,
refreshToken: string,
options: AxiosRequestConfig = {}
) {
const {
accessToken,
userId
} = await acquireToken(refreshToken);
logger.info(`url: ${uri}`);
const result = await axios({
method,
url: `${BASE_URL}${uri}`,
params: options.params,
data: options.data,
headers: { headers: {
Authorization: `Bearer ${accessToken}`, Authorization: `Bearer ${accessToken}`,
'X-Traffic-Id': userId, Referer: 'https://kimi.moonshot.cn/',
...FAKE_HEADERS, 'X-Traffic-Id': `7${util.generateRandomString({ length: 18, charset: 'numeric' })}`,
...(options.headers || {}) ...FAKE_HEADERS
}, },
timeout: options.timeout || 15000, timeout: 15000,
responseType: options.responseType,
validateStatus: () => true validateStatus: () => true
}); });
return checkResult(result, refreshToken); return checkResult(result, refreshToken);
@ -176,17 +153,27 @@ export async function request(
* *
* @param refreshToken access_token的refresh_token * @param refreshToken access_token的refresh_token
*/ */
async function createConversation(model: string, name: string, refreshToken: string) { async function createConversation(name: string, refreshToken: string) {
const {
accessToken,
userId
} = await acquireToken(refreshToken);
const result = await axios.post('https://kimi.moonshot.cn/api/chat', {
is_example: false,
name
}, {
headers: {
Authorization: `Bearer ${accessToken}`,
Referer: 'https://kimi.moonshot.cn/',
'X-Traffic-Id': userId,
...FAKE_HEADERS
},
timeout: 15000,
validateStatus: () => true
});
const { const {
id: convId id: convId
} = await request('POST', '/api/chat', refreshToken, { } = checkResult(result, refreshToken);
data: {
enter_method: 'new_chat',
is_example: false,
kimiplus_id: /^[0-9a-z]{20}$/.test(model) ? model : 'kimi',
name
}
});
return convId; return convId;
} }
@ -198,77 +185,49 @@ async function createConversation(model: string, name: string, refreshToken: str
* @param refreshToken access_token的refresh_token * @param refreshToken access_token的refresh_token
*/ */
async function removeConversation(convId: string, refreshToken: string) { async function removeConversation(convId: string, refreshToken: string) {
return await request('DELETE', `/api/chat/${convId}`, refreshToken); const {
} accessToken,
userId
/** } = await acquireToken(refreshToken);
* const result = await axios.delete(`https://kimi.moonshot.cn/api/chat/${convId}`, {
* headers: {
* @param refreshToken access_token的refresh_token Authorization: `Bearer ${accessToken}`,
*/ Referer: `https://kimi.moonshot.cn/chat/${convId}`,
async function getSuggestion(query: string, refreshToken: string) { 'X-Traffic-Id': userId,
return await request('POST', '/api/suggestion', refreshToken, { ...FAKE_HEADERS
data: { },
offset: 0, timeout: 15000,
page_referer: 'chat', validateStatus: () => true
query: query.replace('user:', '').replace('assistant:', ''),
scene: 'first_round',
size: 10
}
}); });
checkResult(result, refreshToken);
} }
/** /**
* N2S * prompt片段提交
* *
* N2S * @param query prompt
*
* @param model
* @param messages gpt系列消息格式
* @param refs ID列表
* @param refreshToken access_token的refresh_token * @param refreshToken access_token的refresh_token
* @param refConvId ID
*/ */
async function preN2s(model: string, messages: { role: string, content: string }[], refs: string[], refreshToken: string, refConvId?: string) { async function promptSnippetSubmit(query: string, refreshToken: string) {
const isSearchModel = model.indexOf('search') != -1; const {
return await request('POST', `/api/chat/${refConvId}/pre-n2s`, refreshToken, { accessToken,
data: { userId
is_pro_search: false, } = await acquireToken(refreshToken);
kimiplus_id: /^[0-9a-z]{20}$/.test(model) ? model : 'kimi', const result = await axios.post('https://kimi.moonshot.cn/api/prompt-snippet/instance', {
messages, "offset": 0,
refs, "size": 10,
use_search: isSearchModel "query": query.replace('user:', '').replace('assistant:', '')
} }, {
headers: {
Authorization: `Bearer ${accessToken}`,
Referer: 'https://kimi.moonshot.cn/',
'X-Traffic-Id': userId,
...FAKE_HEADERS
},
timeout: 15000,
validateStatus: () => true
}); });
} checkResult(result, refreshToken);
/**
* token计数
*
* @param query
* @param refreshToken access_token的refresh_token
* @param refConvId ID
*/
async function tokenSize(query: string, refs: string[], refreshToken: string, refConvId: string) {
return await request('POST', `/api/chat/${refConvId}/token_size`, refreshToken, {
data: {
content: query,
refs: []
}
});
}
/**
* 使
*
* @param refreshToken access_token的refresh_token
*/
async function getResearchUsage(refreshToken: string): Promise<{
remain,
total,
used
}> {
return await request('GET', '/api/chat/research/usage', refreshToken);
} }
/** /**
@ -277,111 +236,65 @@ async function getResearchUsage(refreshToken: string): Promise<{
* @param model * @param model
* @param messages gpt系列消息格式 * @param messages gpt系列消息格式
* @param refreshToken access_token的refresh_token * @param refreshToken access_token的refresh_token
* @param useSearch
* @param refConvId ID * @param refConvId ID
* @param retryCount * @param retryCount
*/ */
async function createCompletion(model = MODEL_NAME, messages: any[], refreshToken: string, refConvId?: string, retryCount = 0, segmentId?: string): Promise<IStreamMessage> { async function createCompletion(model = MODEL_NAME, messages: any[], refreshToken: string, useSearch = true, refConvId?: string, retryCount = 0) {
return (async () => { return (async () => {
logger.info(messages); logger.info(messages);
// 创建会话
const convId = /[0-9a-zA-Z]{20}/.test(refConvId) ? refConvId : await createConversation(model, "未命名会话", refreshToken);
// 提取引用文件URL并上传kimi获得引用的文件ID列表 // 提取引用文件URL并上传kimi获得引用的文件ID列表
const refFileUrls = extractRefFileUrls(messages); const refFileUrls = extractRefFileUrls(messages);
const refResults = refFileUrls.length ? await Promise.all(refFileUrls.map(fileUrl => uploadFile(fileUrl, refreshToken, convId))) : []; const refs = refFileUrls.length ? await Promise.all(refFileUrls.map(fileUrl => uploadFile(fileUrl, refreshToken))) : [];
const refs = refResults.map(result => result.id);
const refsFile = refResults.map(result => ({
detail: result,
done: true,
file: {},
file_info: result,
id: result.id,
name: result.name,
parse_status: 'success',
size: result.size,
upload_progress: 100,
upload_status: 'success'
}));
// 伪装调用获取用户信息 // 伪装调用获取用户信息
fakeRequest(refreshToken) fakeRequest(refreshToken)
.catch(err => logger.error(err)); .catch(err => logger.error(err));
// 消息预处理 // 创建会话
const convId = /[0-9a-zA-Z]{20}/.test(refConvId) ? refConvId : await createConversation("未命名会话", refreshToken);
// 请求流
const {
accessToken,
userId
} = await acquireToken(refreshToken);
const sendMessages = messagesPrepare(messages, !!refConvId); const sendMessages = messagesPrepare(messages, !!refConvId);
console.log(convId, {
!segmentId && preN2s(model, sendMessages, refs, refreshToken, convId) messages: sendMessages,
.catch(err => logger.error(err)); refs,
getSuggestion(sendMessages[0].content, refreshToken) use_search: useSearch
.catch(err => logger.error(err)); });
tokenSize(sendMessages[0].content, refs, refreshToken, convId) const result = await axios.post(`https://kimi.moonshot.cn/api/chat/${convId}/completion/stream`, {
.catch(err => logger.error(err)); messages: sendMessages,
refs,
const isMath = model.indexOf('math') != -1; use_search: useSearch
const isSearchModel = model.indexOf('search') != -1; }, {
const isResearchModel = model.indexOf('research') != -1;
const isK1Model = model.indexOf('k1') != -1;
logger.info(`使用模型: ${model},是否联网检索: ${isSearchModel},是否探索版: ${isResearchModel}是否K1模型: ${isK1Model},是否数学模型: ${isMath}`);
if(segmentId)
logger.info(`继续请求segmentId: ${segmentId}`);
// 检查探索版使用量
if(isResearchModel) {
const {
total,
used
} = await getResearchUsage(refreshToken);
if(used >= total)
throw new APIException(EX.API_RESEARCH_EXCEEDS_LIMIT, `探索版使用量已达到上限`);
logger.info(`探索版当前额度: ${used}/${total}`);
}
const kimiplusId = isK1Model ? 'crm40ee9e5jvhsn7ptcg' : (/^[0-9a-z]{20}$/.test(model) ? model : 'kimi');
// 请求补全流
const stream = await request('POST', `/api/chat/${convId}/completion/stream`, refreshToken, {
data: segmentId ? {
segment_id: segmentId,
action: 'continue',
messages: [{ role: 'user', content: ' ' }],
kimiplus_id: kimiplusId,
extend: { sidebar: true }
} : {
kimiplus_id: kimiplusId,
messages: sendMessages,
refs,
refs_file: refsFile,
use_math: isMath,
use_research: isResearchModel,
use_search: isSearchModel,
extend: { sidebar: true }
},
headers: { headers: {
Referer: `https://kimi.moonshot.cn/chat/${convId}` Authorization: `Bearer ${accessToken}`,
Referer: `https://kimi.moonshot.cn/chat/${convId}`,
'Priority': 'u=1, i',
'X-Traffic-Id': userId,
...FAKE_HEADERS
}, },
// 120秒超时
timeout: 120000,
validateStatus: () => true,
responseType: 'stream' responseType: 'stream'
}); });
const streamStartTime = util.timestamp(); const streamStartTime = util.timestamp();
// 接收流为输出文本 // 接收流为输出文本
const answer = await receiveStream(model, convId, stream); const answer = await receiveStream(model, convId, result.data);
// 如果上次请求生成长度超限,则继续请求
if(answer.choices[0].finish_reason == 'length' && answer.segment_id) {
const continueAnswer = await createCompletion(model, [], refreshToken, convId, retryCount, answer.segment_id);
answer.choices[0].message.content += continueAnswer.choices[0].message.content;
}
logger.success(`Stream has completed transfer ${util.timestamp() - streamStartTime}ms`); logger.success(`Stream has completed transfer ${util.timestamp() - streamStartTime}ms`);
// 异步移除会话,如果消息不合规,此操作可能会抛出数据库错误异常,请忽略 // 异步移除会话,如果消息不合规,此操作可能会抛出数据库错误异常,请忽略
// 如果引用会话将不会清除,因为我们不知道什么时候你会结束会话 // 如果引用会话将不会清除,因为我们不知道什么时候你会结束会话
!refConvId && removeConversation(convId, refreshToken) !refConvId && removeConversation(convId, refreshToken)
.catch(err => console.error(err)); .catch(err => console.error(err));
promptSnippetSubmit(sendMessages[0].content, refreshToken)
.catch(err => console.error(err));
return answer; return answer;
})() })()
@ -391,7 +304,7 @@ async function createCompletion(model = MODEL_NAME, messages: any[], refreshToke
logger.warn(`Try again after ${RETRY_DELAY / 1000}s...`); logger.warn(`Try again after ${RETRY_DELAY / 1000}s...`);
return (async () => { return (async () => {
await new Promise(resolve => setTimeout(resolve, RETRY_DELAY)); await new Promise(resolve => setTimeout(resolve, RETRY_DELAY));
return createCompletion(model, messages, refreshToken, refConvId, retryCount + 1); return createCompletion(model, messages, refreshToken, useSearch, refConvId, retryCount + 1);
})(); })();
} }
throw err; throw err;
@ -404,92 +317,58 @@ async function createCompletion(model = MODEL_NAME, messages: any[], refreshToke
* @param model * @param model
* @param messages gpt系列消息格式 * @param messages gpt系列消息格式
* @param refreshToken access_token的refresh_token * @param refreshToken access_token的refresh_token
* @param useSearch
* @param refConvId ID * @param refConvId ID
* @param retryCount * @param retryCount
*/ */
async function createCompletionStream(model = MODEL_NAME, messages: any[], refreshToken: string, refConvId?: string, retryCount = 0) { async function createCompletionStream(model = MODEL_NAME, messages: any[], refreshToken: string, useSearch = true, refConvId?: string, retryCount = 0) {
return (async () => { return (async () => {
logger.info(messages); logger.info(messages);
// 创建会话
const convId = /[0-9a-zA-Z]{20}/.test(refConvId) ? refConvId : await createConversation(model, "未命名会话", refreshToken);
// 提取引用文件URL并上传kimi获得引用的文件ID列表 // 提取引用文件URL并上传kimi获得引用的文件ID列表
const refFileUrls = extractRefFileUrls(messages); const refFileUrls = extractRefFileUrls(messages);
const refResults = refFileUrls.length ? await Promise.all(refFileUrls.map(fileUrl => uploadFile(fileUrl, refreshToken, convId))) : []; const refs = refFileUrls.length ? await Promise.all(refFileUrls.map(fileUrl => uploadFile(fileUrl, refreshToken))) : [];
const refs = refResults.map(result => result.id);
const refsFile = refResults.map(result => ({
detail: result,
done: true,
file: {},
file_info: result,
id: result.id,
name: result.name,
parse_status: 'success',
size: result.size,
upload_progress: 100,
upload_status: 'success'
}));
// 伪装调用获取用户信息 // 伪装调用获取用户信息
fakeRequest(refreshToken) fakeRequest(refreshToken)
.catch(err => logger.error(err)); .catch(err => logger.error(err));
// 创建会话
const convId = /[0-9a-zA-Z]{20}/.test(refConvId) ? refConvId : await createConversation("未命名会话", refreshToken);
// 请求流
const {
accessToken,
userId
} = await acquireToken(refreshToken);
const sendMessages = messagesPrepare(messages, !!refConvId); const sendMessages = messagesPrepare(messages, !!refConvId);
const result = await axios.post(`https://kimi.moonshot.cn/api/chat/${convId}/completion/stream`, {
preN2s(model, sendMessages, refs, refreshToken, convId) messages: sendMessages,
.catch(err => logger.error(err)); refs,
getSuggestion(sendMessages[0].content, refreshToken) use_search: useSearch
.catch(err => logger.error(err)); }, {
tokenSize(sendMessages[0].content, refs, refreshToken, convId) // 120秒超时
.catch(err => logger.error(err)); timeout: 120000,
const isMath = model.indexOf('math') != -1;
const isSearchModel = model.indexOf('search') != -1;
const isResearchModel = model.indexOf('research') != -1;
const isK1Model = model.indexOf('k1') != -1;
logger.info(`使用模型: ${model},是否联网检索: ${isSearchModel},是否探索版: ${isResearchModel}是否K1模型: ${isK1Model},是否数学模型: ${isMath}`);
// 检查探索版使用量
if(isResearchModel) {
const {
total,
used
} = await getResearchUsage(refreshToken);
if(used >= total)
throw new APIException(EX.API_RESEARCH_EXCEEDS_LIMIT, `探索版使用量已达到上限`);
logger.info(`探索版当前额度: ${used}/${total}`);
}
const kimiplusId = isK1Model ? 'crm40ee9e5jvhsn7ptcg' : (/^[0-9a-z]{20}$/.test(model) ? model : 'kimi');
// 请求补全流
const stream = await request('POST', `/api/chat/${convId}/completion/stream`, refreshToken, {
data: {
kimiplus_id: kimiplusId,
messages: sendMessages,
refs,
refs_file: refsFile,
use_math: isMath,
use_research: isResearchModel,
use_search: isSearchModel,
extend: { sidebar: true }
},
headers: { headers: {
Referer: `https://kimi.moonshot.cn/chat/${convId}` Authorization: `Bearer ${accessToken}`,
Referer: `https://kimi.moonshot.cn/chat/${convId}`,
'Priority': 'u=1, i',
'X-Traffic-Id': userId,
...FAKE_HEADERS
}, },
validateStatus: () => true,
responseType: 'stream' responseType: 'stream'
}); });
const streamStartTime = util.timestamp(); const streamStartTime = util.timestamp();
// 创建转换流将消息格式转换为gpt兼容格式 // 创建转换流将消息格式转换为gpt兼容格式
return createTransStream(model, convId, stream, () => { return createTransStream(model, convId, result.data, () => {
logger.success(`Stream has completed transfer ${util.timestamp() - streamStartTime}ms`); logger.success(`Stream has completed transfer ${util.timestamp() - streamStartTime}ms`);
// 流传输结束后异步移除会话,如果消息不合规,此操作可能会抛出数据库错误异常,请忽略 // 流传输结束后异步移除会话,如果消息不合规,此操作可能会抛出数据库错误异常,请忽略
// 如果引用会话将不会清除,因为我们不知道什么时候你会结束会话 // 如果引用会话将不会清除,因为我们不知道什么时候你会结束会话
!refConvId && removeConversation(convId, refreshToken) !refConvId && removeConversation(convId, refreshToken)
.catch(err => console.error(err)); .catch(err => console.error(err));
promptSnippetSubmit(sendMessages[0].content, refreshToken)
.catch(err => console.error(err));
}); });
})() })()
.catch(err => { .catch(err => {
@ -498,7 +377,7 @@ async function createCompletionStream(model = MODEL_NAME, messages: any[], refre
logger.warn(`Try again after ${RETRY_DELAY / 1000}s...`); logger.warn(`Try again after ${RETRY_DELAY / 1000}s...`);
return (async () => { return (async () => {
await new Promise(resolve => setTimeout(resolve, RETRY_DELAY)); await new Promise(resolve => setTimeout(resolve, RETRY_DELAY));
return createCompletionStream(model, messages, refreshToken, refConvId, retryCount + 1); return createCompletionStream(model, messages, refreshToken, useSearch, refConvId, retryCount + 1);
})(); })();
} }
throw err; throw err;
@ -513,29 +392,32 @@ async function createCompletionStream(model = MODEL_NAME, messages: any[], refre
* @param refreshToken access_token的refresh_token * @param refreshToken access_token的refresh_token
*/ */
async function fakeRequest(refreshToken: string) { async function fakeRequest(refreshToken: string) {
const {
accessToken,
userId
} = await acquireToken(refreshToken);
const options = {
headers: {
Authorization: `Bearer ${accessToken}`,
Referer: `https://kimi.moonshot.cn/`,
'X-Traffic-Id': userId,
...FAKE_HEADERS
}
};
await [ await [
() => request('GET', '/api/user', refreshToken), () => axios.get('https://kimi.moonshot.cn/api/user', options),
() => request('POST', '/api/user/usage', refreshToken, { () => axios.get('https://kimi.moonshot.cn/api/chat_1m/user/status', options),
data: { () => axios.post('https://kimi.moonshot.cn/api/chat/list', {
usage: ['kimiv', 'math'] offset: 0,
} size: 50
}), }, options),
() => request('GET', '/api/chat_1m/user/status', refreshToken), () => axios.post('https://kimi.moonshot.cn/api/show_case/list', {
() => request('GET', '/api/kimi_mv/user/status', refreshToken), offset: 0,
() => request('POST', '/api/kimiplus/history', refreshToken), size: 4,
() => request('POST', '/api/kimiplus/search', refreshToken, { enable_cache: true,
data: { order: "asc"
offset: 0, }, options)
size: 20 ][Math.floor(Math.random() * 4)]();
}
}),
() => request('POST', '/api/chat/list', refreshToken, {
data: {
offset: 0,
size: 50
}
}),
][Math.floor(Math.random() * 7)]();
} }
/** /**
@ -579,25 +461,12 @@ function extractRefFileUrls(messages: any[]) {
* @param isRefConv * @param isRefConv
*/ */
function messagesPrepare(messages: any[], isRefConv = false) { function messagesPrepare(messages: any[], isRefConv = false) {
let content; // 注入消息提升注意力
if (isRefConv || messages.length < 2) { let latestMessage = messages[messages.length - 1];
content = messages.reduce((content, message) => { let hasFileOrImage = Array.isArray(latestMessage.content)
if (_.isArray(message.content)) { && latestMessage.content.some(v => (typeof v === 'object' && ['file', 'image_url'].includes(v['type'])));
return message.content.reduce((_content, v) => { // 第二轮开始注入system prompt
if (!_.isObject(v) || v['type'] != 'text') return _content; if (messages.length > 2) {
return _content + `${v["text"] || ""}\n`;
}, content);
}
return content += `${message.role == 'user' ? wrapUrlsToTags(message.content) : message.content}\n`;
}, '')
logger.info("\n透传内容\n" + content);
}
else {
// 注入消息提升注意力
let latestMessage = messages[messages.length - 1];
let hasFileOrImage = Array.isArray(latestMessage.content)
&& latestMessage.content.some(v => (typeof v === 'object' && ['file', 'image_url'].includes(v['type'])));
// 第二轮开始注入system prompt
if (hasFileOrImage) { if (hasFileOrImage) {
let newFileMessage = { let newFileMessage = {
"content": "关注用户最新发送文件和消息", "content": "关注用户最新发送文件和消息",
@ -613,6 +482,22 @@ function messagesPrepare(messages: any[], isRefConv = false) {
messages.splice(messages.length - 1, 0, newTextMessage); messages.splice(messages.length - 1, 0, newTextMessage);
logger.info("注入提升尾部消息注意力system prompt"); logger.info("注入提升尾部消息注意力system prompt");
} }
}
let content;
if (isRefConv || messages.length < 2) {
content = messages.reduce((content, message) => {
if (_.isArray(message.content)) {
return message.content.reduce((_content, v) => {
if (!_.isObject(v) || v['type'] != 'text') return _content;
return _content + `${v["text"] || ""}\n`;
}, content);
}
return content += `${message.role == 'user' ? wrapUrlsToTags(message.content) : message.content}\n`;
}, '')
logger.info("\n透传内容\n" + content);
}
else {
content = messages.reduce((content, message) => { content = messages.reduce((content, message) => {
if (_.isArray(message.content)) { if (_.isArray(message.content)) {
return message.content.reduce((_content, v) => { return message.content.reduce((_content, v) => {
@ -647,13 +532,13 @@ function wrapUrlsToTags(content: string) {
* @param filename * @param filename
* @param refreshToken access_token的refresh_token * @param refreshToken access_token的refresh_token
*/ */
async function preSignUrl(action: string, filename: string, refreshToken: string) { async function preSignUrl(filename: string, refreshToken: string) {
const { const {
accessToken, accessToken,
userId userId
} = await acquireToken(refreshToken); } = await acquireToken(refreshToken);
const result = await axios.post('https://kimi.moonshot.cn/api/pre-sign-url', { const result = await axios.post('https://kimi.moonshot.cn/api/pre-sign-url', {
action, action: 'file',
name: filename name: filename
}, { }, {
timeout: 15000, timeout: 15000,
@ -695,9 +580,8 @@ async function checkFileUrl(fileUrl: string) {
* *
* @param fileUrl URL * @param fileUrl URL
* @param refreshToken access_token的refresh_token * @param refreshToken access_token的refresh_token
* @param refConvId ID
*/ */
async function uploadFile(fileUrl: string, refreshToken: string, refConvId?: string) { async function uploadFile(fileUrl: string, refreshToken: string) {
// 预检查远程文件URL可用性 // 预检查远程文件URL可用性
await checkFileUrl(fileUrl); await checkFileUrl(fileUrl);
@ -721,14 +605,11 @@ async function uploadFile(fileUrl: string, refreshToken: string, refConvId?: str
})); }));
} }
const fileType = (mimeType || '').includes('image') ? 'image' : 'file';
// 获取预签名文件URL // 获取预签名文件URL
let { const {
url: uploadUrl, url: uploadUrl,
object_name: objectName, object_name: objectName
file_id: fileId } = await preSignUrl(filename, refreshToken);
} = await preSignUrl(fileType, filename, refreshToken);
// 获取文件的MIME类型 // 获取文件的MIME类型
mimeType = mimeType || mime.getType(filename); mimeType = mimeType || mime.getType(filename);
@ -756,22 +637,16 @@ async function uploadFile(fileUrl: string, refreshToken: string, refConvId?: str
}); });
checkResult(result, refreshToken); checkResult(result, refreshToken);
let status, startTime = Date.now(); let fileId, status, startTime = Date.now();
let fileDetail; while (status != 'initialized') {
while (status != 'initialized' && status != 'parsed') {
if (Date.now() - startTime > 30000) if (Date.now() - startTime > 30000)
throw new Error('文件等待处理超时'); throw new Error('文件等待处理超时');
// 获取文件上传结果 // 获取文件上传结果
result = await axios.post('https://kimi.moonshot.cn/api/file', fileType == 'image' ? { result = await axios.post('https://kimi.moonshot.cn/api/file', {
type: 'image',
file_id: fileId,
name: filename
} : {
type: 'file', type: 'file',
name: filename, name: filename,
object_name: objectName, object_name: objectName,
file_id: '', timeout: 15000
chat_id: refConvId
}, { }, {
headers: { headers: {
Authorization: `Bearer ${accessToken}`, Authorization: `Bearer ${accessToken}`,
@ -780,12 +655,11 @@ async function uploadFile(fileUrl: string, refreshToken: string, refConvId?: str
...FAKE_HEADERS ...FAKE_HEADERS
} }
}); });
fileDetail = checkResult(result, refreshToken); ({ id: fileId, status } = checkResult(result, refreshToken));
({ id: fileId, status } = fileDetail);
} }
startTime = Date.now(); startTime = Date.now();
let parseFinish = status == 'parsed'; let parseFinish = false;
while (!parseFinish) { while (!parseFinish) {
if (Date.now() - startTime > 30000) if (Date.now() - startTime > 30000)
throw new Error('文件等待处理超时'); throw new Error('文件等待处理超时');
@ -807,7 +681,7 @@ async function uploadFile(fileUrl: string, refreshToken: string, refConvId?: str
}); });
} }
return fileDetail; return fileId;
} }
/** /**
@ -840,9 +714,7 @@ function checkResult(result: AxiosResponse, refreshToken: string) {
* @param convId ID * @param convId ID
* @param stream * @param stream
*/ */
async function receiveStream(model: string, convId: string, stream: any): Promise<IStreamMessage> { async function receiveStream(model: string, convId: string, stream: any) {
let webSearchCount = 0;
let temp = Buffer.from('');
return new Promise((resolve, reject) => { return new Promise((resolve, reject) => {
// 消息初始化 // 消息初始化
const data = { const data = {
@ -853,11 +725,10 @@ async function receiveStream(model: string, convId: string, stream: any): Promis
{ index: 0, message: { role: 'assistant', content: '' }, finish_reason: 'stop' } { index: 0, message: { role: 'assistant', content: '' }, finish_reason: 'stop' }
], ],
usage: { prompt_tokens: 1, completion_tokens: 1, total_tokens: 2 }, usage: { prompt_tokens: 1, completion_tokens: 1, total_tokens: 2 },
segment_id: '',
created: util.unixTimestamp() created: util.unixTimestamp()
}; };
let refContent = ''; let refContent = '';
const silentSearch = model.indexOf('silent') != -1; const silentSearch = model.indexOf('silent_search') != -1;
const parser = createParser(event => { const parser = createParser(event => {
try { try {
if (event.type !== "event") return; if (event.type !== "event") return;
@ -867,16 +738,8 @@ async function receiveStream(model: string, convId: string, stream: any): Promis
throw new Error(`Stream response invalid: ${event.data}`); throw new Error(`Stream response invalid: ${event.data}`);
// 处理消息 // 处理消息
if (result.event == 'cmpl' && result.text) { if (result.event == 'cmpl' && result.text) {
data.choices[0].message.content += result.text; const exceptCharIndex = result.text.indexOf("<22>");
} data.choices[0].message.content += result.text.substring(0, exceptCharIndex == -1 ? result.text.length : exceptCharIndex);
// 处理请求ID
else if(result.event == 'req') {
data.segment_id = result.id;
}
// 处理超长文本
else if(result.event == 'length') {
logger.warn('此次生成达到max_tokens稍候将继续请求拼接完整响应');
data.choices[0].finish_reason = 'length';
} }
// 处理结束或错误 // 处理结束或错误
else if (result.event == 'all_done' || result.event == 'error') { else if (result.event == 'all_done' || result.event == 'error') {
@ -885,10 +748,8 @@ async function receiveStream(model: string, convId: string, stream: any): Promis
resolve(data); resolve(data);
} }
// 处理联网搜索 // 处理联网搜索
else if (!silentSearch && result.event == 'search_plus' && result.msg && result.msg.type == 'get_res') { else if (!silentSearch && result.event == 'search_plus' && result.msg && result.msg.type == 'get_res')
webSearchCount += 1; refContent += `${result.msg.title} - ${result.msg.url}\n`;
refContent += `【检索 ${webSearchCount}】 [${result.msg.title}](${result.msg.url})\n\n`;
}
// else // else
// logger.warn(result.event, result); // logger.warn(result.event, result);
} }
@ -898,20 +759,7 @@ async function receiveStream(model: string, convId: string, stream: any): Promis
} }
}); });
// 将流数据喂给SSE转换器 // 将流数据喂给SSE转换器
stream.on("data", buffer => { stream.on("data", buffer => parser.feed(buffer.toString()));
// 检查buffer是否以完整UTF8字符结尾
if (buffer.toString().indexOf('<27>') != -1) {
// 如果不完整则累积buffer直到收到完整字符
temp = Buffer.concat([temp, buffer]);
return;
}
// 将之前累积的不完整buffer拼接
if (temp.length > 0) {
buffer = Buffer.concat([temp, buffer]);
temp = Buffer.from('');
}
parser.feed(buffer.toString());
});
stream.once("error", err => reject(err)); stream.once("error", err => reject(err));
stream.once("close", () => resolve(data)); stream.once("close", () => resolve(data));
}); });
@ -932,11 +780,8 @@ function createTransStream(model: string, convId: string, stream: any, endCallba
const created = util.unixTimestamp(); const created = util.unixTimestamp();
// 创建转换流 // 创建转换流
const transStream = new PassThrough(); const transStream = new PassThrough();
let webSearchCount = 0;
let searchFlag = false; let searchFlag = false;
let lengthExceed = false; const silentSearch = model.indexOf('silent_search') != -1;
let segmentId = '';
const silentSearch = model.indexOf('silent') != -1;
!transStream.closed && transStream.write(`data: ${JSON.stringify({ !transStream.closed && transStream.write(`data: ${JSON.stringify({
id: convId, id: convId,
model, model,
@ -944,7 +789,6 @@ function createTransStream(model: string, convId: string, stream: any, endCallba
choices: [ choices: [
{ index: 0, delta: { role: 'assistant', content: '' }, finish_reason: null } { index: 0, delta: { role: 'assistant', content: '' }, finish_reason: null }
], ],
segment_id: '',
created created
})}\n\n`); })}\n\n`);
const parser = createParser(event => { const parser = createParser(event => {
@ -965,21 +809,12 @@ function createTransStream(model: string, convId: string, stream: any, endCallba
choices: [ choices: [
{ index: 0, delta: { content: (searchFlag ? '\n' : '') + chunk }, finish_reason: null } { index: 0, delta: { content: (searchFlag ? '\n' : '') + chunk }, finish_reason: null }
], ],
segment_id: segmentId,
created created
})}\n\n`; })}\n\n`;
if (searchFlag) if (searchFlag)
searchFlag = false; searchFlag = false;
!transStream.closed && transStream.write(data); !transStream.closed && transStream.write(data);
} }
// 处理请求ID
else if(result.event == 'req') {
segmentId = result.id;
}
// 处理超长文本
else if (result.event == 'length') {
lengthExceed = true;
}
// 处理结束或错误 // 处理结束或错误
else if (result.event == 'all_done' || result.event == 'error') { else if (result.event == 'all_done' || result.event == 'error') {
const data = `data: ${JSON.stringify({ const data = `data: ${JSON.stringify({
@ -990,11 +825,10 @@ function createTransStream(model: string, convId: string, stream: any, endCallba
{ {
index: 0, delta: result.event == 'error' ? { index: 0, delta: result.event == 'error' ? {
content: '\n[内容由于不合规被停止生成,我们换个话题吧]' content: '\n[内容由于不合规被停止生成,我们换个话题吧]'
} : {}, finish_reason: lengthExceed ? 'length' : 'stop' } : {}, finish_reason: 'stop'
} }
], ],
usage: { prompt_tokens: 1, completion_tokens: 1, total_tokens: 2 }, usage: { prompt_tokens: 1, completion_tokens: 1, total_tokens: 2 },
segment_id: segmentId,
created created
})}\n\n`; })}\n\n`;
!transStream.closed && transStream.write(data); !transStream.closed && transStream.write(data);
@ -1005,7 +839,6 @@ function createTransStream(model: string, convId: string, stream: any, endCallba
else if (!silentSearch && result.event == 'search_plus' && result.msg && result.msg.type == 'get_res') { else if (!silentSearch && result.event == 'search_plus' && result.msg && result.msg.type == 'get_res') {
if (!searchFlag) if (!searchFlag)
searchFlag = true; searchFlag = true;
webSearchCount += 1;
const data = `data: ${JSON.stringify({ const data = `data: ${JSON.stringify({
id: convId, id: convId,
model, model,
@ -1013,11 +846,10 @@ function createTransStream(model: string, convId: string, stream: any, endCallba
choices: [ choices: [
{ {
index: 0, delta: { index: 0, delta: {
content: `检索 ${webSearchCount}】 [${result.msg.title}](${result.msg.url})\n` content: `检索 ${result.msg.title} - ${result.msg.url} ...\n`
}, finish_reason: null }, finish_reason: null
} }
], ],
segment_id: segmentId,
created created
})}\n\n`; })}\n\n`;
!transStream.closed && transStream.write(data); !transStream.closed && transStream.write(data);

View File

@ -1,20 +0,0 @@
export default interface IStreamMessage {
id: string;
model: string;
object: string;
choices: {
index: number;
message: {
role: string;
content: string;
};
finish_reason: string;
}[];
usage: {
prompt_tokens: number;
completion_tokens: number;
total_tokens: number;
};
segment_id?: string;
created: number;
}

View File

@ -20,19 +20,15 @@ export default {
const tokens = chat.tokenSplit(request.headers.authorization); const tokens = chat.tokenSplit(request.headers.authorization);
// 随机挑选一个refresh_token // 随机挑选一个refresh_token
const token = _.sample(tokens); const token = _.sample(tokens);
let { model, conversation_id: convId, messages, stream, use_search } = request.body; const { model, conversation_id: convId, messages, stream, use_search } = request.body;
if(use_search)
model = 'kimi-search';
if (stream) { if (stream) {
const stream = await chat.createCompletionStream(model, messages, token, convId); const stream = await chat.createCompletionStream(model, messages, token, use_search, convId);
return new Response(stream, { return new Response(stream, {
type: "text/event-stream" type: "text/event-stream"
}); });
} }
else else
return await chat.createCompletion(model, messages, token, convId); return await chat.createCompletion(model, messages, token, use_search, convId);
} }
} }

1555
yarn.lock

File diff suppressed because it is too large Load Diff