Update main.go
Browse files
main.go
CHANGED
@@ -143,6 +143,7 @@ func handleModels(c *gin.Context) {
|
|
143 |
{"id": "claude-3-haiku", "object": "model", "owned_by": "ddg"},
|
144 |
{"id": "llama-3.1-70b", "object": "model", "owned_by": "ddg"},
|
145 |
{"id": "mixtral-8x7b", "object": "model", "owned_by": "ddg"},
|
|
|
146 |
}
|
147 |
c.JSON(http.StatusOK, gin.H{"object": "list", "data": models})
|
148 |
}
|
@@ -347,42 +348,7 @@ func handleCompletion(c *gin.Context) {
|
|
347 |
}
|
348 |
}
|
349 |
|
350 |
-
|
351 |
-
// req, err := http.NewRequest("GET", "https://duckduckgo.com/duckchat/v1/status", nil)
|
352 |
-
// if err != nil {
|
353 |
-
// return "", fmt.Errorf("创建请求失败: %v", err)
|
354 |
-
// }
|
355 |
-
// for k, v := range config.FakeHeaders {
|
356 |
-
// req.Header.Set(k, v)
|
357 |
-
// }
|
358 |
-
// req.Header.Set("x-vqd-accept", "1")
|
359 |
-
//
|
360 |
-
// client := &http.Client{
|
361 |
-
// Timeout: 10 * time.Second,
|
362 |
-
// }
|
363 |
-
//
|
364 |
-
// log.Println("发送 token 请求")
|
365 |
-
// resp, err := client.Do(req)
|
366 |
-
// if err != nil {
|
367 |
-
// return "", fmt.Errorf("请求失败: %v", err)
|
368 |
-
// }
|
369 |
-
// defer resp.Body.Close()
|
370 |
-
//
|
371 |
-
// if resp.StatusCode != http.StatusOK {
|
372 |
-
// bodyBytes, _ := io.ReadAll(resp.Body)
|
373 |
-
// bodyString := string(bodyBytes)
|
374 |
-
// log.Printf("requestToken: 非200响应: %d, 内容: %s\n", resp.StatusCode, bodyString)
|
375 |
-
// return "", fmt.Errorf("非200响应: %d, 内容: %s", resp.StatusCode, bodyString)
|
376 |
-
// }
|
377 |
-
//
|
378 |
-
// token := resp.Header.Get("x-vqd-4")
|
379 |
-
// if token == "" {
|
380 |
-
// return "", errors.New("响应中未包含x-vqd-4头")
|
381 |
-
// }
|
382 |
-
//
|
383 |
-
// // log.Printf("获取到的 token: %s\n", token)
|
384 |
-
// return token, nil
|
385 |
-
//}
|
386 |
|
387 |
func requestToken() (string, error) {
|
388 |
url := "https://duckduckgo.com/duckchat/v1/status"
|
@@ -493,6 +459,8 @@ func convertModel(inputModel string) string {
|
|
493 |
return "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo"
|
494 |
case "mixtral-8x7b":
|
495 |
return "mistralai/Mixtral-8x7B-Instruct-v0.1"
|
|
|
|
|
496 |
default:
|
497 |
return "gpt-4o-mini"
|
498 |
}
|
|
|
143 |
{"id": "claude-3-haiku", "object": "model", "owned_by": "ddg"},
|
144 |
{"id": "llama-3.1-70b", "object": "model", "owned_by": "ddg"},
|
145 |
{"id": "mixtral-8x7b", "object": "model", "owned_by": "ddg"},
|
146 |
+
{"id": "o3-mini", "object": "model", "owned_by": "ddg"},
|
147 |
}
|
148 |
c.JSON(http.StatusOK, gin.H{"object": "list", "data": models})
|
149 |
}
|
|
|
348 |
}
|
349 |
}
|
350 |
|
351 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
352 |
|
353 |
func requestToken() (string, error) {
|
354 |
url := "https://duckduckgo.com/duckchat/v1/status"
|
|
|
459 |
return "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo"
|
460 |
case "mixtral-8x7b":
|
461 |
return "mistralai/Mixtral-8x7B-Instruct-v0.1"
|
462 |
+
case "o3-mini":
|
463 |
+
return "o3-mini"
|
464 |
default:
|
465 |
return "gpt-4o-mini"
|
466 |
}
|