ai/llama3.3

Verified Publisher

By Docker

Updated 10 months ago

Newest LLama 3 release with improved reasoning and generation quality

Model
18

50K+

ai/llama3.3 repository overview

Llama 3.3

logo

Meta Llama 3.3 is a powerful 70B parameter multilingual language model designed by Meta for text-based tasks like chat and content generation. The instruction-tuned version is optimized for multilingual dialogue and performs better than many open-source and commercial models on common benchmarks.

Intended uses

  • Multilingual assistant-like chat: Using instruction-tuned models for conversational AI across multiple languages, enabling natural and context-aware interactions in various linguistic settings.

  • Coding support and software development tasks: Leveraging language models to assist with code generation, debugging, documentation, and other software engineering workflows.

  • Multilingual content creation and localization: Generating and adapting written content across different languages and cultures, supporting global communication and engagement.

  • Knowledge-based applications: Integrating LLMs with structured or unstructured data sources to answer questions, extract insights, or support decision-making.

  • General natural language generation: Various NLG tasks such as summarization, translation, or content generation across different domains.

  • Synthetic data generation (synth): Creating realistic, high-quality synthetic text data to augment datasets for training, testing, or anonymization purposes.

Characteristics

AttributeDetails
ProviderMeta
Architecturellama
Cutoff dateDecember 2023
LanguagesEnglish, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.
Tool calling
Input modalitiesText
Output modalitiesText and Code
LicenseLlama 3.3 Community license

Available model variants

Model variantParametersQuantizationContext windowVRAM¹Size
ai/llama3.3:latest

ai/llama3.3:70B-Q4_K_M
70BIQ2_XXS/Q4_K_M131K tokens41.11 GiB39.59 GB
ai/llama3.3:70B-Q4_070BQ4_0131K tokens38.73 GiB37.22 GB
ai/llama3.3:70B-Q4_K_M70BIQ2_XXS/Q4_K_M131K tokens41.11 GiB39.59 GB

¹: VRAM estimated based on model characteristics.

latest70B-Q4_K_M

Use this AI model with Docker Model Runner

First, pull the model:

docker model pull ai/llama3.3

Then run the model:

docker model run ai/llama3.3

For more information on Docker Model Runner, explore the documentation.

Benchmark performance

CategoryBenchmarkLlama-3.3 70B Instruct
GeneralMMLU (CoT)86.0
MMLU Pro (CoT)68.9
SteerabilityIFEval92.1
ReasoningGPQA Diamond (CoT)50.5
CodeHumanEval88.4
MBPP EvalPlus (base)87.6
MathMATH (CoT)77.0
Tool UseBFCL v277.3
MultilingualMGSM91.1

Tag summary

Content type

Model

Digest

sha256:c6c3d506c

Size

37.2 GB

Last updated

10 months ago

docker model pull ai/llama3.3:70B-Q4_0

This week's pulls

Pulls:

1,561

Last week