Cohere Transcribe 03-2026 โ€” MLX 8-bit

This repository contains an MLX-native int8 conversion of Cohere Transcribe 03-2026 for local automatic speech recognition on Apple Silicon.

Note This repo is a community mirror of the canonical MLX conversion maintained by AppAutomaton at appautomaton/cohere-asr-mlx.

Variants

Path Precision
mlx-int8/ int8 quantized weights

Model Details

  • Developed by: AppAutomaton
  • Shared by: mlx-community
  • Original MLX repo: appautomaton/cohere-asr-mlx
  • Upstream model: CohereLabs/cohere-transcribe-03-2026
  • Task: automatic speech recognition
  • Runtime: MLX on Apple Silicon

How to Get Started

Command-line transcription with mlx-speech:

python scripts/generate/cohere_asr.py \
  --audio input.wav \
  --output transcript.txt

Minimal Python usage:

import numpy as np
import soundfile as sf

from mlx_speech.generation import CohereAsrModel

audio, sr = sf.read("input.wav", dtype="float32", always_2d=False)
if audio.ndim > 1:
    audio = audio.mean(axis=1)
if sr != 16000:
    old_len = len(audio)
    new_len = int(round(old_len * 16000 / sr))
    audio = np.interp(
        np.linspace(0, old_len - 1, new_len),
        np.arange(old_len),
        audio,
    ).astype(np.float32)

model = CohereAsrModel.from_path("mlx-int8")
result = model.transcribe(audio, sample_rate=16000, language="en")
print(result.text)

Notes

  • This repo contains the quantized MLX runtime artifact only.
  • The conversion keeps the original encoder-decoder ASR architecture and remaps weights explicitly for MLX inference.
  • The example above resamples to 16 kHz before calling transcribe(), which matches the runtime requirement.
  • This mirror is a duplicated repo, not an automatically synchronized namespace mirror.

Links

License

Apache 2.0 โ€” following the upstream Cohere Transcribe model license. Check the original Cohere release for current terms.

Downloads last month

-

Downloads are not tracked for this model. How to track
MLX
Hardware compatibility
Log In to add your hardware

Quantized

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for mlx-community/cohere-transcribe-03-2026-mlx-8bit

Quantized
(24)
this model