license: mit tags: - bitnet - ternary - trillim - cpu-inference base_model: 1bitLLM/bitnet_b1_58-large
Ternary-quantized version of 1bitLLM/bitnet_b1_58-large, packaged for the Trillim DarkNet inference engine.
This model runs entirely on CPU — no GPU required.
pip install trillim trillim pull Trillim/BitNet-Large-TRNQ trillim serve Trillim/BitNet-Large-TRNQ
This starts an OpenAI-compatible API server at http://127.0.0.1:8000.
http://127.0.0.1:8000
For interactive CLI chat:
trillim chat Trillim/BitNet-Large-TRNQ
qmodel.tensors
rope.cache
config.json
tokenizer.json
tokenizer_config.json
tokenizer.model
tokenization_bitnet.py
trillim_config.json
This model is released under the MIT License, following the license of the source model.