Llama 3.3 70B
Latest open-weight model from Meta with significant improvements in reasoning and instruction following. Strong general-purpose performance.
Quick Overview
Pricing
Simple, transparent token-based pricing. Pay only for what you use.
Capabilities
Core strengths and supported features of Llama 3.3 70B.
Performance
Context window capacity and inference speed characteristics.
Context Window
Standard context window that handles most use cases including document analysis, code review, and lengthy conversations.
Speed: Fast
Optimized for low-latency responses. Ideal for real-time applications, chatbots, and interactive experiences where response time is critical.
Quick Start
Get started with Llama 3.3 70B in just a few lines of code.
import OriginalPoint from "originalpoint";
const client = new OriginalPoint({
apiKey: process.env.ORIGINALPOINT_API_KEY,
});
const response = await client.chat.completions.create({
model: "llama-3.3-70b",
messages: [
{
role: "user",
content: "Hello, how can you help me today?",
},
],
max_tokens: 1024,
});
console.log(response.choices[0].message.content);More from Meta
Explore other models available from Meta.
Llama 3.1 70B
BalancedProven open-weight model with excellent multilingual support and strong coding abilities. Popular choice for cost-effective deployments.
Llama 3.1 8B
EconomyLightweight open model optimized for edge deployment and high throughput. Great for simple tasks, embeddings, and cost-sensitive applications.