DevelopersLearnCompanyPricing
Sign In

Runtime Benchmarks

Compare LLM endpoints with live performance benchmarks

Documentation

Learn how to use the Unify API

Showcase

Explore LLM apps built by our community using Unify

Blog

Read about LLM deployment infrastructure

Paper Readings

Join our discussions around cuttin-edge AI research

Talks

Dive deep with us into the AI landscape

Careers

Join our team and let’s Unify AI!

Contact

Reach out to our team

Privacy & Cookies

How we treat your navigation data

Terms Of Service

General requirements for using our Service

Team

Learn more about the people behind Unify

Socials

Follow us through our social accounts:
Sign In
Developers

Runtime Benchmarks

Compare LLM endpoints with live performance benchmarks

Documentation

Learn how to use the Unify API

Showcase

Explore LLM apps built by our community using Unify

Learn

Blog

Read about LLM deployment infrastructure

Paper Readings

Join our discussions around cuttin-edge AI research

Talks

Dive deep with us into the AI landscape

Company

Careers

Join our team and let’s Unify AI!

Contact

Reach out to our team

Privacy & Cookies

How we treat your navigation data

Terms Of Service

General requirements for using our Service

Team

Learn more about the people behind Unify

Socials

Follow us through our social accounts:
Pricing

Benchmarks

Learn More
Compare the performance of LLMs across endpoint providers to find the best possible configuration for your speed, latency and cost requirements. Our objective benchmarks are continuously updated with the newest models and endpoints.

Models

Select from the list of models below

mixtral-8x7b-instruct-v0.1
llama-3.1-8b-chat
llama-3.1-70b-chat
llama-3.1-405b-chat
llama-3-8b-chat
llama-3-70b-chat
mistral-7b-instruct-v0.2
mistral-nemo
mistral-7b-instruct-v0.3
gemma-2-9b-it
llama-3.2-3b-chat
codellama-34b-instruct
gemma-7b-it
mixtral-8x22b-instruct-v0.1
llama-3.2-11b-chat
llama-3.2-90b-chat
mistral-large
qwen-2-72b-instruct
llama-3.2-1b-chat
mistral-7b-instruct-v0.1
claude-3-haiku
claude-3-opus
claude-3-sonnet
claude-3.5-sonnet
qwen-2.5-72b-instruct
claude-3.5-sonnet-20240620
gpt-4o
gpt-4o-mini
gpt-4o-2024-08-06
gemma-2-27b-it
1
2
Developers
Runtime BenchmarksDocumentationApps Showcase
Learn
BlogPaper ReadingsTalks
Socials
DiscordLinkedInTwitterYouTube
Company
CareersContactPrivacy PolicyTerms Of ServiceTeam