Verified models.
AI & ML interests
None defined yet.
Recent Activity
View all activity
Verified models. Tested with vLLM.
LLMs quantized with LLM Compressor to NVFP4. Check the JSON files in the model directories for more information.
-
kaitchup/QwQ-32B-bnb-4bit
Text Generation • 34B • Updated • 13 • 1 -
kaitchup/QwQ-32B-AutoRoundGPTQ-8bit
Text Generation • 33B • Updated • 64 • 2 -
kaitchup/QwQ-32B-AutoRoundGPTQ-4bit
Text Generation • 33B • Updated • 11 • 1 -
kaitchup/QwQ-32B-AutoRoundGPTQ-3bit
Text Generation • 31B • Updated • 12
-
kaitchup/OLMo-2-1124-7B-Instruct-AutoRound-GPTQ-4bit
Text Generation • 7B • Updated • 12 • 1 -
kaitchup/Llama-3.1-Tulu-3-70B-AutoRound-GPTQ-4bit
Text Generation • 71B • Updated • 9 -
kaitchup/Llama-3.1-Tulu-3-8B-AutoRound-GPTQ-4bit
Text Generation • 8B • Updated • 4 -
kaitchup/OLMo-2-1124-13B-Instruct-AutoRound-GPTQ-4bit
Text Generation • 14B • Updated • 10
Some language pairs of OPUS formatted to have source and target sentences as single sequences. Intended to facilitate fine-tuning of causal LLMs.
Llama 2 7B, 13B, Llama 3 8B, and Mistral 7B quantized with GPTQ in 2-bit, 3-bit, 4-bit and 8-bit with GPTQ.
-
kaitchup/Llama-2-7b-hf-gptq-8bit
Text Generation • 7B • Updated • 5 -
kaitchup/Llama-2-7b-hf-gptq-3bit
Text Generation • 6B • Updated • 879 • 1 -
kaitchup/Llama-2-7b-hf-gptq-2bit
Text Generation • 7B • Updated • 868 -
kaitchup/Llama-2-7b-hf-gptq-4bit
Text Generation • 7B • Updated • 7
Contaminated Mistral 7B and TinyLlama adapters, and the datasets used for contamination.
-
kaitchup/Mistral-NeMo-Minitron-8B-Base-AutoRound-GPTQ-sym-4bit
Text Generation • 8B • Updated • 5 -
kaitchup/Mistral-NeMo-Minitron-8B-Base-AutoRound-GPTQ-asym-4bit
Text Generation • 8B • Updated • 5 -
kaitchup/Mistral-Nemo-Base-2407-AutoRound-GPTQ-asym-4bit
Text Generation • 12B • Updated • 11 -
kaitchup/Llama-3.1-Minitron-4B-Width-Base-AutoRound-GPTQ-asym-4bit
Text Generation • 5B • Updated • 6
Verified models.
Verified models. All compatible with vLLM for very fast inference. Use the 3.1 models as they are more recent.
-
kaitchup/Qwen2.5-72B-Instruct-autoround-2bit-32g-4096-gptq
73B • Updated • 8 -
kaitchup/Qwen2.5-72B-Instruct-autoround-2bit-128g-4096-gptq
73B • Updated • 7 -
kaitchup/Qwen2.5-72B-Instruct-autoround-2bit-128g-2048-gptq
73B • Updated • 6 -
kaitchup/Qwen2.5-72B-Instruct-autoround-2bit-64g-4096-gptq
73B • Updated • 6
-
kaitchup/Phi-4-mini-instruct-AutoRoundGPTQ-8bit
Text Generation • 4B • Updated • 7 -
kaitchup/Phi-4-mini-instruct-AutoRoundGPTQ-3bit
Text Generation • 4B • Updated • 9 -
kaitchup/Phi-4-mini-instruct-AutoRoundGPTQ-4bit
Text Generation • 4B • Updated • 10 -
kaitchup/Phi-4-mini-instruct-AutoRoundGPTQ-2bit
Text Generation • 4B • Updated • 8
-
kaitchup/DeepSeek-R1-Distill-Llama-8B-AutoRound-GPTQ-4bit
Text Generation • 8B • Updated • 382 • 1 -
kaitchup/DeepSeek-R1-Distill-Qwen-14B-AutoRound-GPTQ-4bit
Text Generation • 15B • Updated • 28 • 7 -
kaitchup/DeepSeek-R1-Distill-Qwen-7B-AutoRound-GPTQ-4bit
Text Generation • 8B • Updated • 26 • 2
-
kaitchup/Falcon3-10B-Instruct-AutoRound-GPTQ-4bit
Text Generation • 10B • Updated • 8 -
kaitchup/Falcon3-10B-Base-AutoRound-GPTQ-4bit
Text Generation • 10B • Updated • 11 -
kaitchup/Falcon3-7B-Base-AutoRound-GPTQ-4bit
Text Generation • 7B • Updated • 10 -
kaitchup/Falcon3-7B-Instruct-AutoRound-GPTQ-4bit
Text Generation • 7B • Updated • 6
-
kaitchup/Qwen2.5-1.5B-AutoRound-GPTQ-asym-4bit
Text Generation • 2B • Updated • 7 -
kaitchup/Qwen2.5-7B-AutoRound-GPTQ-asym-4bit
Text Generation • 8B • Updated • 4 -
kaitchup/Qwen2.5-1.5B-Instruct-AutoRound-GPTQ-asym-4bit
Text Generation • 2B • Updated • 7 -
kaitchup/Qwen2.5-7B-Instruct-AutoRound-GPTQ-asym-4bit
Text Generation • 8B • Updated • 7
Machine translation adapters for Llama 2 7B.
-
kaitchup/Llama-2-7b-mt-French-to-English
Translation • Updated • 135 • 3 -
kaitchup/Llama-2-7b-mt-Italian-to-English
Translation • Updated • 7 -
kaitchup/Llama-2-7b-mt-Indonesian-to-English
Translation • Updated • 12 -
kaitchup/Llama-2-7b-mt-Vietnamese-to-English
Translation • Updated • 9
Quantized and fine-tuned versions of the Yi models
A collection of 7B models made with mergekit.
-
kaitchup/Phi-3-mini-4k-instruct-gptq-4bit
Text Generation • 4B • Updated • 733k • 2 -
kaitchup/Phi-3-medium-128k-instruct-awq-4bit
Text Generation • 14B • Updated -
kaitchup/Phi-3-mini-4k-instruct-bnb-4bit
Text Generation • 4B • Updated • 7 -
kaitchup/Phi-3-medium-4k-instruct-awq-4bit
Text Generation • 14B • Updated
-
kaitchup/Meta-Llama-3.1-8B-Instruct-autoround-gptq-4bit-sym
Text Generation • 8B • Updated • 8 • 1 -
kaitchup/Meta-Llama-3.1-8B-Instruct-awq-4bit
Text Generation • 8B • Updated • 14 • 1 -
kaitchup/Meta-Llama-3.1-8B-awq-4bit
Text Generation • 8B • Updated • 7 -
kaitchup/Meta-Llama-3.1-8B-Instruct-gptq-4bit
Text Generation • 8B • Updated • 7
-
kaitchup/Mistral-NeMo-Minitron-8B-Base-Minivoc-32k-v0.1a
Text Generation • 8B • Updated • 2 -
kaitchup/Llama-3.1-8B-Minivoc-32k-v0.1a
Text Generation • 7B • Updated • 1 -
kaitchup/Qwen2-1.5B-Minivoc-32k-v0.1a
Text Generation • 1B • Updated • 12 • 2 -
kaitchup/Qwen2.5-1.5B-Minivoc-32k-v0.1a-AutoRound-GPTQ-asym-4bit
Text Generation • 1B • Updated • 6
Verified models.
Verified models.
Verified models. Tested with vLLM.
Verified models. All compatible with vLLM for very fast inference. Use the 3.1 models as they are more recent.
LLMs quantized with LLM Compressor to NVFP4. Check the JSON files in the model directories for more information.
-
kaitchup/Qwen2.5-72B-Instruct-autoround-2bit-32g-4096-gptq
73B • Updated • 8 -
kaitchup/Qwen2.5-72B-Instruct-autoround-2bit-128g-4096-gptq
73B • Updated • 7 -
kaitchup/Qwen2.5-72B-Instruct-autoround-2bit-128g-2048-gptq
73B • Updated • 6 -
kaitchup/Qwen2.5-72B-Instruct-autoround-2bit-64g-4096-gptq
73B • Updated • 6
-
kaitchup/QwQ-32B-bnb-4bit
Text Generation • 34B • Updated • 13 • 1 -
kaitchup/QwQ-32B-AutoRoundGPTQ-8bit
Text Generation • 33B • Updated • 64 • 2 -
kaitchup/QwQ-32B-AutoRoundGPTQ-4bit
Text Generation • 33B • Updated • 11 • 1 -
kaitchup/QwQ-32B-AutoRoundGPTQ-3bit
Text Generation • 31B • Updated • 12
-
kaitchup/Phi-4-mini-instruct-AutoRoundGPTQ-8bit
Text Generation • 4B • Updated • 7 -
kaitchup/Phi-4-mini-instruct-AutoRoundGPTQ-3bit
Text Generation • 4B • Updated • 9 -
kaitchup/Phi-4-mini-instruct-AutoRoundGPTQ-4bit
Text Generation • 4B • Updated • 10 -
kaitchup/Phi-4-mini-instruct-AutoRoundGPTQ-2bit
Text Generation • 4B • Updated • 8
-
kaitchup/DeepSeek-R1-Distill-Llama-8B-AutoRound-GPTQ-4bit
Text Generation • 8B • Updated • 382 • 1 -
kaitchup/DeepSeek-R1-Distill-Qwen-14B-AutoRound-GPTQ-4bit
Text Generation • 15B • Updated • 28 • 7 -
kaitchup/DeepSeek-R1-Distill-Qwen-7B-AutoRound-GPTQ-4bit
Text Generation • 8B • Updated • 26 • 2
-
kaitchup/Falcon3-10B-Instruct-AutoRound-GPTQ-4bit
Text Generation • 10B • Updated • 8 -
kaitchup/Falcon3-10B-Base-AutoRound-GPTQ-4bit
Text Generation • 10B • Updated • 11 -
kaitchup/Falcon3-7B-Base-AutoRound-GPTQ-4bit
Text Generation • 7B • Updated • 10 -
kaitchup/Falcon3-7B-Instruct-AutoRound-GPTQ-4bit
Text Generation • 7B • Updated • 6
-
kaitchup/OLMo-2-1124-7B-Instruct-AutoRound-GPTQ-4bit
Text Generation • 7B • Updated • 12 • 1 -
kaitchup/Llama-3.1-Tulu-3-70B-AutoRound-GPTQ-4bit
Text Generation • 71B • Updated • 9 -
kaitchup/Llama-3.1-Tulu-3-8B-AutoRound-GPTQ-4bit
Text Generation • 8B • Updated • 4 -
kaitchup/OLMo-2-1124-13B-Instruct-AutoRound-GPTQ-4bit
Text Generation • 14B • Updated • 10
-
kaitchup/Qwen2.5-1.5B-AutoRound-GPTQ-asym-4bit
Text Generation • 2B • Updated • 7 -
kaitchup/Qwen2.5-7B-AutoRound-GPTQ-asym-4bit
Text Generation • 8B • Updated • 4 -
kaitchup/Qwen2.5-1.5B-Instruct-AutoRound-GPTQ-asym-4bit
Text Generation • 2B • Updated • 7 -
kaitchup/Qwen2.5-7B-Instruct-AutoRound-GPTQ-asym-4bit
Text Generation • 8B • Updated • 7
Some language pairs of OPUS formatted to have source and target sentences as single sequences. Intended to facilitate fine-tuning of causal LLMs.
Machine translation adapters for Llama 2 7B.
-
kaitchup/Llama-2-7b-mt-French-to-English
Translation • Updated • 135 • 3 -
kaitchup/Llama-2-7b-mt-Italian-to-English
Translation • Updated • 7 -
kaitchup/Llama-2-7b-mt-Indonesian-to-English
Translation • Updated • 12 -
kaitchup/Llama-2-7b-mt-Vietnamese-to-English
Translation • Updated • 9
Llama 2 7B, 13B, Llama 3 8B, and Mistral 7B quantized with GPTQ in 2-bit, 3-bit, 4-bit and 8-bit with GPTQ.
-
kaitchup/Llama-2-7b-hf-gptq-8bit
Text Generation • 7B • Updated • 5 -
kaitchup/Llama-2-7b-hf-gptq-3bit
Text Generation • 6B • Updated • 879 • 1 -
kaitchup/Llama-2-7b-hf-gptq-2bit
Text Generation • 7B • Updated • 868 -
kaitchup/Llama-2-7b-hf-gptq-4bit
Text Generation • 7B • Updated • 7
Quantized and fine-tuned versions of the Yi models
Contaminated Mistral 7B and TinyLlama adapters, and the datasets used for contamination.
A collection of 7B models made with mergekit.
-
kaitchup/Phi-3-mini-4k-instruct-gptq-4bit
Text Generation • 4B • Updated • 733k • 2 -
kaitchup/Phi-3-medium-128k-instruct-awq-4bit
Text Generation • 14B • Updated -
kaitchup/Phi-3-mini-4k-instruct-bnb-4bit
Text Generation • 4B • Updated • 7 -
kaitchup/Phi-3-medium-4k-instruct-awq-4bit
Text Generation • 14B • Updated
-
kaitchup/Meta-Llama-3.1-8B-Instruct-autoround-gptq-4bit-sym
Text Generation • 8B • Updated • 8 • 1 -
kaitchup/Meta-Llama-3.1-8B-Instruct-awq-4bit
Text Generation • 8B • Updated • 14 • 1 -
kaitchup/Meta-Llama-3.1-8B-awq-4bit
Text Generation • 8B • Updated • 7 -
kaitchup/Meta-Llama-3.1-8B-Instruct-gptq-4bit
Text Generation • 8B • Updated • 7
-
kaitchup/Mistral-NeMo-Minitron-8B-Base-AutoRound-GPTQ-sym-4bit
Text Generation • 8B • Updated • 5 -
kaitchup/Mistral-NeMo-Minitron-8B-Base-AutoRound-GPTQ-asym-4bit
Text Generation • 8B • Updated • 5 -
kaitchup/Mistral-Nemo-Base-2407-AutoRound-GPTQ-asym-4bit
Text Generation • 12B • Updated • 11 -
kaitchup/Llama-3.1-Minitron-4B-Width-Base-AutoRound-GPTQ-asym-4bit
Text Generation • 5B • Updated • 6
-
kaitchup/Mistral-NeMo-Minitron-8B-Base-Minivoc-32k-v0.1a
Text Generation • 8B • Updated • 2 -
kaitchup/Llama-3.1-8B-Minivoc-32k-v0.1a
Text Generation • 7B • Updated • 1 -
kaitchup/Qwen2-1.5B-Minivoc-32k-v0.1a
Text Generation • 1B • Updated • 12 • 2 -
kaitchup/Qwen2.5-1.5B-Minivoc-32k-v0.1a-AutoRound-GPTQ-asym-4bit
Text Generation • 1B • Updated • 6