vocaela-for-edge
Collection
5 items • Updated
This repo contains GGUF-format weights for Vocaela-500M, containing:
Note: So far LlamaCpp has problem on rendering chat template correctly. To workaround it, we apply chat template (e.g., using python / node.js etc.) before calling llama-server endpoint. For examples of how to use it, please refer to repo vocaela-500m-demo
8-bit
Base model
HuggingFaceTB/SmolLM2-360M