MetaBalancedNew

Llama 3.3 70B

Latest open-weight model from Meta with significant improvements in reasoning and instruction following. Strong general-purpose performance.

4.5(18,900 ratings)

Quick Overview

CategoryGeneral Purpose
Context Window128K tokens
SpeedFast
Input$0.60 / 1M tokens
Output$0.60 / 1M tokens

Pricing

Simple, transparent token-based pricing. Pay only for what you use.

Input Tokens
$0.60
per 1M tokens
Output Tokens
$0.60
per 1M tokens

Capabilities

Core strengths and supported features of Llama 3.3 70B.

Text Generation
Code
Analysis

Performance

Context window capacity and inference speed characteristics.

Context Window

128K

Standard context window that handles most use cases including document analysis, code review, and lengthy conversations.

Speed: Fast

Optimized for low-latency responses. Ideal for real-time applications, chatbots, and interactive experiences where response time is critical.

Quick Start

Get started with Llama 3.3 70B in just a few lines of code.

import OriginalPoint from "originalpoint";

const client = new OriginalPoint({
  apiKey: process.env.ORIGINALPOINT_API_KEY,
});

const response = await client.chat.completions.create({
  model: "llama-3.3-70b",
  messages: [
    {
      role: "user",
      content: "Hello, how can you help me today?",
    },
  ],
  max_tokens: 1024,
});

console.log(response.choices[0].message.content);