Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings
Discussion options

What if you could run a fast quantized LLM and get performance similar to an unquantized model requiring four times the resources? That's exactly what the new NexaQuant models deliver.

Based on the popular DeepSeek R1, the NexaQuant distilled models have reasoning capabilities close to, or in some cases even exceeding!, their unquantized R1-distill sources.

Get started now with Nexa SDK:

DeepSeek R1 Distill Llama 8B: nexa run DeepSeek-R1-Distill-Llama-8B-NexaQuant:q4_0

DeepSeek R1 Distill Qwen 1.5B: nexa run DeepSeek-R1-Distill-Qwen-1.5B-NexaQuant:q4_0

* Disclaimer: I'm not affiliated with Nexa AI. Information presented is unofficial and subject to change without notice.

You must be logged in to vote

Replies: 0 comments

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
1 participant
Morty Proxy This is a proxified and sanitized view of the page, visit original site.