TypeGPTProxy / main.py
AIMaster7's picture
Update main.py
ef4717a verified
from fastapi import FastAPI, Request, Header, HTTPException
from fastapi.responses import JSONResponse, Response, StreamingResponse
import httpx
import socket
app = FastAPI()
# Your real TypeGPT API key
REAL_API_KEY = "sk-kwWVUQPDsLemvilLcyzSSWRzo8sctCzxzlbdN0ZC5ZUCCv0m"
BASE_URL = "https://fast.typegpt.net"
PUBLIC_AUTH_TOKEN = "TypeGPT-Free4ALL"
@app.on_event("startup")
async def startup_event():
try:
hostname = socket.gethostname()
server_ip = socket.gethostbyname(hostname)
print("===== Server Started =====")
print(f"πŸ“‘ Server IP: {server_ip}")
except Exception as e:
print(f"⚠️ Could not determine server IP: {e}")
@app.api_route("/{path:path}", methods=["GET", "POST"])
async def proxy(request: Request, path: str, authorization: str = Header(None)):
# Validate Authorization header
if not authorization or not authorization.startswith("Bearer "):
raise HTTPException(status_code=401, detail="Missing or malformed Authorization header.")
token = authorization.replace("Bearer ", "").strip()
if token != PUBLIC_AUTH_TOKEN:
raise HTTPException(status_code=401, detail="Invalid Authorization token. Use 'TypeGPT-Free4ALL'.")
# Build backend request
target_url = f"{BASE_URL}/{path}"
headers = {
"Authorization": f"Bearer {REAL_API_KEY}",
"Content-Type": request.headers.get("content-type", "application/json"),
"Accept": "text/event-stream",
"User-Agent": "FastAPI-Proxy"
}
body = await request.body()
is_stream = b'"stream":true' in body or b'"stream": true' in body
print(f"πŸ”„ Forwarding to: {target_url}")
print(f"πŸ“¦ Request Body (first 200 chars): {body[:200]}")
print(f"πŸ” Stream mode: {is_stream}")
if is_stream:
# Stream response immediately, as chunks arrive
async def stream_generator():
async with httpx.AsyncClient(timeout=None) as client:
async with client.stream(
method=request.method,
url=target_url,
headers=headers,
content=body
) as upstream_response:
async for chunk in upstream_response.aiter_raw():
yield chunk # raw bytes streamed directly
return StreamingResponse(
stream_generator(),
status_code=200,
media_type="text/event-stream"
)
else:
# Normal JSON response
async with httpx.AsyncClient(timeout=60) as client:
try:
response = await client.request(
method=request.method,
url=target_url,
headers=headers,
content=body
)
except httpx.RequestError as e:
raise HTTPException(status_code=502, detail=f"Backend request failed: {e}")
print(f"↩️ TypeGPT Status: {response.status_code}")
print(f"🧾 Response Snippet: {response.text[:200]}")
try:
return JSONResponse(content=response.json(), status_code=response.status_code)
except Exception:
return Response(
content=response.content,
status_code=response.status_code,
media_type=response.headers.get("content-type", "text/plain")
)