Performance Metrics

The following tables present comprehensive benchmarking results across different models and evaluation criteria.

Average Numeric Scores

ModelAverage Scores (%)Token CountResponse Time (s)Factual Accuracy (%)Hallucination Rate (%)Logical Correctness (%)Source Quality (%)
Claude87162310484889181
Gemini921050160091929588
Manus79427889173729069
OpenAI93931673793939589
Perplexity88315218086899382
xAI Grok 3 Deep Search8929293983899386
xAI Grok 3 Deeper Search85290522980809278

Qualitative Ratings

ModelRatingToken CountResponse TimeFactual AccuracyMistakesCorrectnessSource Quality
Claude3253443
Gemini4525555
Manus1311141
OpenAI4525555
Perplexity3254544
xAI Grok 3 Deep Search3253445
xAI Grok 3 Deeper Search2242243

Subscription Costs

ModelCost
OpenAI ChatGPT Deep Research$200/mo (Pro, 250 queries/mo); $20/mo (Plus, Team, Edu, Enterprise, 25 queries/mo)
Anthropic Claude Research$100 or $200/mo (Max tier only, US/JP/BR, early beta)
Gemini 2.5 Pro Deep Research$20/month
Perplexity Deep ResearchFree (5 queries/day) or $20/month
xAI Grok 3 Deep Research$30/month (SuperGrok)
xAI Grok 3 Deeper Research$30/month (SuperGrok - Deeper Mode)
Manus AI$2-10 per task (depending on task intensity & difficulty)