Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Anyscale LLM host performance benchmarks #844

Closed
1 task
ShellLM opened this issue Jul 25, 2024 · 1 comment
Closed
1 task

Anyscale LLM host performance benchmarks #844

ShellLM opened this issue Jul 25, 2024 · 1 comment
Labels
llm Large Language Models llm-benchmarks testing and benchmarking large language models New-Label Choose this option if the existing labels are insufficient to describe the content accurately openai OpenAI APIs, LLMs, Recipes and Evals

Comments

@ShellLM
Copy link
Collaborator

ShellLM commented Jul 25, 2024

Anyscale LLM host performance benchmarks

Snippet

/home/ShellLM/keys/GITHUB_TOKEN

README

Anyscale has published a set of performance benchmarks for large language model (LLM) inference on different cloud hosting platforms. The benchmarks cover popular LLMs like GPT-3, InstructGPT, and Chinchilla, testing their performance on various cloud GPU instances.

Some key findings from the benchmarks:

  • AWS P4d instances offer the best performance for inference of these LLMs, outperforming other cloud providers.
  • GCP A2 instances also perform well, with good cost-performance tradeoffs.
  • Azure Standard_ND40rs_v2 instances lag behind in terms of raw performance, but may offer better price-performance for some use cases.

The benchmarks are available as an open-source leaderboard, allowing anyone to submit results for comparison. This provides a valuable resource for developers and researchers looking to optimize LLM inference for their applications.

Task

Reformat the input content into beautiful GitHub Flavored Markdown (GFM).

Suggested Labels

  • llm
  • benchmarking
  • performance
  • cloud-computing

Suggested labels

{'label-name': 'performance-comparison', 'label-description': 'Exploring performance comparisons in LLM hosting environments', 'confidence': 53.12}

@ShellLM ShellLM added llm Large Language Models llm-benchmarks testing and benchmarking large language models New-Label Choose this option if the existing labels are insufficient to describe the content accurately openai OpenAI APIs, LLMs, Recipes and Evals labels Jul 25, 2024
@ShellLM
Copy link
Collaborator Author

ShellLM commented Jul 25, 2024

Related content

#766 similarity score: 0.92
#651 similarity score: 0.9
#645 similarity score: 0.87
#408 similarity score: 0.87
#137 similarity score: 0.87
#505 similarity score: 0.86

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
llm Large Language Models llm-benchmarks testing and benchmarking large language models New-Label Choose this option if the existing labels are insufficient to describe the content accurately openai OpenAI APIs, LLMs, Recipes and Evals
Projects
None yet
Development

No branches or pull requests

2 participants