Safetensors
GGUF
English
llama
conversational
How to use from the
Use from the
llama-cpp-python library
# !pip install llama-cpp-python

from llama_cpp import Llama

llm = Llama.from_pretrained(
	repo_id="QKing-Official/EndAI-Small",
	filename="",
)
llm.create_chat_completion(
	messages = "No input example has been defined for this model task."
)

EndAI‑small (TinyLlama)

This is a small AI model. It's build on top of TinyLlama and trained on 3% of the HuggingFaceH4/ultrachat_200k dataset. This model is designed to be small and able to run on cpu and gpu really quickly.

Downloads last month
227
Safetensors
Model size
1B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for QKing-Official/EndAI-Small

Quantized
(145)
this model

Dataset used to train QKing-Official/EndAI-Small