Spaces:
Running
Running
Update main.py
Browse files
main.py
CHANGED
@@ -48,12 +48,25 @@ async def proxy(request: Request, path: str, target_url: Optional[str] = None):
|
|
48 |
# 移除可能导致问题的头
|
49 |
headers.pop("host", None)
|
50 |
headers.pop("content-length", None)
|
51 |
-
|
|
|
|
|
52 |
# 检查是否请求流式响应
|
53 |
stream_request = "stream" in request.query_params and request.query_params["stream"].lower() in ["true", "1", "yes"]
|
54 |
|
55 |
# 创建cloudscraper实例
|
56 |
-
scraper = cloudscraper.create_scraper(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
57 |
|
58 |
# 获取请求体
|
59 |
body = await request.body()
|
|
|
48 |
# 移除可能导致问题的头
|
49 |
headers.pop("host", None)
|
50 |
headers.pop("content-length", None)
|
51 |
+
headers.update({
|
52 |
+
"sec-fetch-dest": "document",
|
53 |
+
})
|
54 |
# 检查是否请求流式响应
|
55 |
stream_request = "stream" in request.query_params and request.query_params["stream"].lower() in ["true", "1", "yes"]
|
56 |
|
57 |
# 创建cloudscraper实例
|
58 |
+
scraper = cloudscraper.create_scraper(
|
59 |
+
browser={
|
60 |
+
'browser': 'chrome',
|
61 |
+
'platform': 'windows',
|
62 |
+
'desktop': True
|
63 |
+
},
|
64 |
+
delay=10
|
65 |
+
)
|
66 |
+
|
67 |
+
scraper.get("https://grok.com/",headers={
|
68 |
+
"sec-fetch-dest": "document",
|
69 |
+
})
|
70 |
|
71 |
# 获取请求体
|
72 |
body = await request.body()
|