g4f-backup / usage /2025-02-20.jsonl
roxky's picture
Save history
254d63b
raw
history blame
1.62 kB
{"model": "r1-1776", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 113, "total_tokens": 139}
{"model": "r1-1776", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 119, "total_tokens": 145}
{"model": "sonar-reasoning", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 220, "total_tokens": 246}
{"model": "r1-1776", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 121, "total_tokens": 147}
{"model": "r1-1776", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 27, "total_tokens": 53}
{"model": "sonar-reasoning-pro", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 569, "total_tokens": 595}
{"model": "sonar-reasoning-pro", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 587, "total_tokens": 613}
{"model": "sonar-reasoning-pro", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 804, "total_tokens": 830}
{"model": "r1-1776", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 143, "total_tokens": 169}
{"model": "r1-1776", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 13, "total_tokens": 39}
{"model": "sonar-reasoning-pro", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 537, "total_tokens": 563}
{"model": "sonar-reasoning-pro", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 678, "total_tokens": 704}
{"model": "sonar-reasoning-pro", "provider": "PerplexityLabs", "prompt_tokens": 26, "completion_tokens": 810, "total_tokens": 836}