--- language: - uk - en license: apache-2.0 library_name: peft tags: - translation - mlx datasets: - Helsinki-NLP/opus_paracrawl - turuta/Multi30k-uk metrics: - bleu pipeline_tag: text-generation base_model: mistralai/Mistral-7B-v0.1 inference: false model-index: - name: Dragoman results: - task: type: translation name: English-Ukrainian Translation dataset: name: FLORES-101 type: facebook/flores config: eng_Latn-ukr_Cyrl split: devtest metrics: - type: bleu value: 32.34 name: Test BLEU --- # lang-uk/dragoman-4bit This model was converted to MLX format from the [`lang-uk/dragoman`](https://huggingface.co/lang-uk/dragoman) adapter fused into the [`mistralai/Mistral-7b-v0.1`](https://huggingface.co/mistralai/Mistral-7B-v0.1) base model and quantized into 4 bits using mlx-lm version **0.4.0**. Refer to the [original model card](https://huggingface.co/lang-uk/dragoman) for more details on the model. ## Use with mlx ```bash pip install mlx-lm ``` ```python from mlx_lm import load, generate model, tokenizer = load("lang-uk/dragoman-4bit") response = generate(model, tokenizer, prompt="[INST] who holds this neighborhood? [/INST]", verbose=True) ``` Or use from your shell: ```console python -m mlx_lm.generate --model lang-uk/dragoman-4bit --prompt '[INST] who holds this neighborhood? [/INST]' --temp 0 --max-tokens 100 ```