GPT-3.5 Turbo: no wait / GPT-4: no wait / GPT-4 Turbo: no wait / Gemini Pro: no wait
{ "uptime": 439169, "endpoints": { "openai": "https://gepettothreepointfive-tnd.hf.space/TND/openai", "openai2": "https://gepettothreepointfive-tnd.hf.space/TND/openai/turbo-instruct", "google-ai": "https://gepettothreepointfive-tnd.hf.space/TND/google-ai" }, "proompts": 7735, "tookens": "23.09m", "proomptersNow": 2, "openaiKeys": 1904, "openaiOrgs": 1802, "google-aiKeys": 58, "turbo": { "usage": "22.55m tokens", "activeKeys": 1095, "revokedKeys": 114, "overQuotaKeys": 695, "trialKeys": 1102, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "gpt4": { "usage": "0 tokens", "activeKeys": 0, "overQuotaKeys": 1, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "gpt4-turbo": { "usage": "0 tokens", "activeKeys": 0, "overQuotaKeys": 1, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "gemini-pro": { "usage": "546.1k tokens", "activeKeys": 58, "revokedKeys": 0, "proomptersInQueue": 0, "estimatedQueueTime": "no wait" }, "config": { "gatekeeper": "proxy_key", "maxIpsAutoBan": "true", "textModelRateLimit": "6", "imageModelRateLimit": "4", "maxContextTokensOpenAI": "16000", "maxContextTokensAnthropic": "200000", "maxOutputTokensOpenAI": "4096", "maxOutputTokensAnthropic": "2048", "allowAwsLogging": "false", "promptLogging": "false", "tokenQuota": { "turbo": "0", "gpt4": "0", "gpt4-32k": "0", "gpt4-turbo": "0", "dall-e": "0", "claude": "0", "claude-opus": "0", "gemini-pro": "0", "mistral-tiny": "0", "mistral-small": "0", "mistral-medium": "0", "mistral-large": "0", "aws-claude": "0", "aws-claude-opus": "0", "azure-turbo": "0", "azure-gpt4": "0", "azure-gpt4-32k": "0", "azure-gpt4-turbo": "0", "azure-dall-e": "0" }, "allowOpenAIToolUsage": "false", "allowImagePrompts": "false" }, "build": "d3e7ef3 (main@khanon/oai-reverse-proxy)" }