@cf/meta/llama-3.3-70b-instruct-fp8-fast vs @hf/thebloke/mistral-7b-instruct-v0.1-awq
Compare @cf/meta/llama-3.3-70b-instruct-fp8-fast by Cloudflare Workers AI and @hf/thebloke/mistral-7b-instruct-v0.1-awq by Cloudflare Workers AI. See detailed specifications, pricing, and capabilities side-by-side.
Quick Overview
@cf/meta/llama-3.3-70b-instruct-fp8-fast
Cloudflare Workers AI
24K tokens context • $0.29 / $2.25 per 1M tokens
View full specifications →@hf/thebloke/mistral-7b-instruct-v0.1-awq
Cloudflare Workers AI
4K tokens context • Free
View full specifications →Detailed Comparison
Specification
@cf/meta/llama-3.3-70b-instruct-fp8-fast
@hf/thebloke/mistral-7b-instruct-v0.1-awq
Provider
Cloudflare Workers AI
Cloudflare Workers AI
Context Window
24K tokens
4K tokens
Max Output Tokens
24K tokens
4K tokens
Input Pricing (per 1M tokens)
$0.29
Free
Output Pricing (per 1M tokens)
$2.25
Free
Release Date
Dec 2024
Sep 2023
Capabilities
Capability
@cf/meta/llama-3.3-70b-instruct-fp8-fast
@hf/thebloke/mistral-7b-instruct-v0.1-awq
Text Generation
Function Calling
Which Model Should You Choose?
Choose @cf/meta/llama-3.3-70b-instruct-fp8-fast if:
- • You need a larger context window
Choose @hf/thebloke/mistral-7b-instruct-v0.1-awq if:
- • Cost efficiency is a priority