mirror of
https://github.com/ChatGPTNextWeb/ChatGPT-Next-Web.git
synced 2025-10-09 19:46:37 +08:00
feat: support https_proxy for standalone mode
This commit is contained in:
parent
2ae62d9c0e
commit
20738d9deb
@ -16,7 +16,7 @@ RUN apk update && apk add --no-cache git
|
|||||||
|
|
||||||
ENV OPENAI_API_KEY=""
|
ENV OPENAI_API_KEY=""
|
||||||
ENV CODE=""
|
ENV CODE=""
|
||||||
ARG DOCKER=true
|
ARG STANDALONE=true
|
||||||
|
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
COPY --from=deps /app/node_modules ./node_modules
|
COPY --from=deps /app/node_modules ./node_modules
|
||||||
|
@ -177,7 +177,13 @@ bash <(curl -s https://raw.githubusercontent.com/Yidadaa/ChatGPT-Next-Web/main/s
|
|||||||
```shell
|
```shell
|
||||||
docker pull yidadaa/chatgpt-next-web
|
docker pull yidadaa/chatgpt-next-web
|
||||||
|
|
||||||
docker run -d -p 3000:3000 -e OPENAI_API_KEY="" -e CODE="" yidadaa/chatgpt-next-web
|
# https_proxy 指向代理,需要带协议前缀 例如: https_proxy=https://192.168.1.1:7890
|
||||||
|
docker run -it --rm \
|
||||||
|
-p 3000:3000 \
|
||||||
|
-e https_proxy="" \
|
||||||
|
-e OPENAI_API_KEY="" \
|
||||||
|
-e CODE="" \
|
||||||
|
--name chatgpt-next-web yidadaa/chatgpt-next-web
|
||||||
```
|
```
|
||||||
|
|
||||||
## 截图 Screenshots
|
## 截图 Screenshots
|
||||||
|
@ -14,4 +14,4 @@ export function getAccessCodes(): Set<string> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
export const ACCESS_CODES = getAccessCodes();
|
export const ACCESS_CODES = getAccessCodes();
|
||||||
export const IS_IN_DOCKER = process.env.DOCKER;
|
export const IS_STANDALONE = process.env.STANDALONE;
|
||||||
|
@ -1,6 +1,7 @@
|
|||||||
import { createParser } from "eventsource-parser";
|
import { createParser } from "eventsource-parser";
|
||||||
import { NextRequest } from "next/server";
|
import { NextRequest } from "next/server";
|
||||||
import { requestOpenai } from "../common";
|
import { requestOpenai } from "../common";
|
||||||
|
import { PageConfig } from "next/types";
|
||||||
|
|
||||||
async function createStream(req: NextRequest) {
|
async function createStream(req: NextRequest) {
|
||||||
const encoder = new TextEncoder();
|
const encoder = new TextEncoder();
|
||||||
@ -56,6 +57,6 @@ export async function POST(req: NextRequest) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
export const config = {
|
export const config: PageConfig = {
|
||||||
runtime: "edge",
|
runtime: process.env.STANDALONE ? "nodejs" : "edge",
|
||||||
};
|
};
|
||||||
|
@ -4,6 +4,9 @@ const OPENAI_URL = "api.openai.com";
|
|||||||
const DEFAULT_PROTOCOL = "https";
|
const DEFAULT_PROTOCOL = "https";
|
||||||
const PROTOCOL = process.env.PROTOCOL ?? DEFAULT_PROTOCOL;
|
const PROTOCOL = process.env.PROTOCOL ?? DEFAULT_PROTOCOL;
|
||||||
const BASE_URL = process.env.BASE_URL ?? OPENAI_URL;
|
const BASE_URL = process.env.BASE_URL ?? OPENAI_URL;
|
||||||
|
const STANDALONE = Boolean(process.env.STANDALONE);
|
||||||
|
|
||||||
|
let fetch: FetchLike = globalThis.fetch;
|
||||||
|
|
||||||
export async function requestOpenai(req: NextRequest) {
|
export async function requestOpenai(req: NextRequest) {
|
||||||
const apiKey = req.headers.get("token");
|
const apiKey = req.headers.get("token");
|
||||||
@ -20,3 +23,64 @@ export async function requestOpenai(req: NextRequest) {
|
|||||||
body: req.body,
|
body: req.body,
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
export type FetchLike = (
|
||||||
|
url: string | Request,
|
||||||
|
init?: RequestInit,
|
||||||
|
) => Promise<Response>;
|
||||||
|
|
||||||
|
if (STANDALONE) {
|
||||||
|
const proxy =
|
||||||
|
process.env.HTTPS_PROXY ||
|
||||||
|
process.env.https_proxy ||
|
||||||
|
process.env.ALL_PROXY ||
|
||||||
|
process.env.all_proxy;
|
||||||
|
if (proxy) {
|
||||||
|
console.log(`[HTTP Proxy] ${new URL(proxy).hostname}`);
|
||||||
|
}
|
||||||
|
|
||||||
|
fetch = createFetchWithProxyByNextUndici({ proxy, fetch });
|
||||||
|
}
|
||||||
|
|
||||||
|
export function createFetchWithProxyByNextUndici({
|
||||||
|
proxy,
|
||||||
|
fetch,
|
||||||
|
}: {
|
||||||
|
proxy?: string;
|
||||||
|
fetch?: FetchLike;
|
||||||
|
} = {}): FetchLike {
|
||||||
|
if (!proxy) {
|
||||||
|
return fetch || globalThis.fetch;
|
||||||
|
}
|
||||||
|
let agent: any;
|
||||||
|
return async (...args) => {
|
||||||
|
const init = (args[1] ||= {});
|
||||||
|
if (init.body instanceof ReadableStream) {
|
||||||
|
// https://github.com/nodejs/node/issues/46221
|
||||||
|
(init as any).duplex ||= "half";
|
||||||
|
}
|
||||||
|
if (!agent) {
|
||||||
|
let ProxyAgent;
|
||||||
|
if ("ProxyAgent" in globalThis) {
|
||||||
|
ProxyAgent = (globalThis as any).ProxyAgent;
|
||||||
|
fetch ||= globalThis.fetch;
|
||||||
|
} else {
|
||||||
|
// @ts-ignore
|
||||||
|
const undici = await import("next/dist/compiled/undici");
|
||||||
|
ProxyAgent = undici.ProxyAgent;
|
||||||
|
fetch ||= undici.fetch;
|
||||||
|
}
|
||||||
|
agent = new ProxyAgent(proxy);
|
||||||
|
// https://github.com/nodejs/node/issues/43187#issuecomment-1134634174
|
||||||
|
(global as any)[Symbol.for("undici.globalDispatcher.1")] = agent;
|
||||||
|
}
|
||||||
|
return fetch!(...args);
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
// @ts-ignore
|
||||||
|
declare module "next/dist/compiled/undici" {
|
||||||
|
const fetch: FetchLike;
|
||||||
|
const ProxyAgent: any;
|
||||||
|
export { fetch, ProxyAgent };
|
||||||
|
}
|
||||||
|
@ -1,5 +1,10 @@
|
|||||||
import { NextRequest, NextResponse } from "next/server";
|
import { NextRequest, NextResponse } from "next/server";
|
||||||
import { requestOpenai } from "../common";
|
import { requestOpenai } from "../common";
|
||||||
|
import { PageConfig } from "next/types";
|
||||||
|
|
||||||
|
export const config: PageConfig = {
|
||||||
|
runtime: process.env.STANDALONE ? "nodejs" : "edge",
|
||||||
|
};
|
||||||
|
|
||||||
async function makeRequest(req: NextRequest) {
|
async function makeRequest(req: NextRequest) {
|
||||||
try {
|
try {
|
||||||
|
@ -3,7 +3,7 @@ import "./styles/globals.scss";
|
|||||||
import "./styles/markdown.scss";
|
import "./styles/markdown.scss";
|
||||||
import "./styles/highlight.scss";
|
import "./styles/highlight.scss";
|
||||||
import process from "child_process";
|
import process from "child_process";
|
||||||
import { ACCESS_CODES, IS_IN_DOCKER } from "./api/access";
|
import { ACCESS_CODES, IS_STANDALONE } from "./api/access";
|
||||||
|
|
||||||
let COMMIT_ID: string | undefined;
|
let COMMIT_ID: string | undefined;
|
||||||
try {
|
try {
|
||||||
@ -29,7 +29,7 @@ export const metadata = {
|
|||||||
function Meta() {
|
function Meta() {
|
||||||
const metas = {
|
const metas = {
|
||||||
version: COMMIT_ID ?? "unknown",
|
version: COMMIT_ID ?? "unknown",
|
||||||
access: ACCESS_CODES.size > 0 || IS_IN_DOCKER ? "enabled" : "disabled",
|
access: ACCESS_CODES.size > 0 || IS_STANDALONE ? "enabled" : "disabled",
|
||||||
};
|
};
|
||||||
|
|
||||||
return (
|
return (
|
||||||
|
@ -1,5 +1,4 @@
|
|||||||
/** @type {import('next').NextConfig} */
|
/** @type {import("next").NextConfig} */
|
||||||
|
|
||||||
const nextConfig = {
|
const nextConfig = {
|
||||||
experimental: {
|
experimental: {
|
||||||
appDir: true,
|
appDir: true,
|
||||||
@ -11,11 +10,11 @@ const nextConfig = {
|
|||||||
}); // 针对 SVG 的处理规则
|
}); // 针对 SVG 的处理规则
|
||||||
|
|
||||||
return config;
|
return config;
|
||||||
}
|
},
|
||||||
};
|
};
|
||||||
|
|
||||||
if (process.env.DOCKER) {
|
if (process.env.STANDALONE) {
|
||||||
nextConfig.output = 'standalone'
|
nextConfig.output = "standalone";
|
||||||
}
|
}
|
||||||
|
|
||||||
module.exports = nextConfig;
|
export default nextConfig;
|
Loading…
Reference in New Issue
Block a user