Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Buckets new
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

aufklarer
/
PersonaPlex-7B-MLX-4bit

Audio-to-Audio
MLX
Moshi
English
personaplex
speech-to-speech
full-duplex
apple-silicon
Model card Files Files and versions
xet
Community

Instructions to use aufklarer/PersonaPlex-7B-MLX-4bit with libraries, inference providers, notebooks, and local apps. Follow these links to get started.

  • Libraries
  • MLX

    How to use aufklarer/PersonaPlex-7B-MLX-4bit with MLX:

    # Download the model from the Hub
    pip install huggingface_hub[hf_xet]
    
    huggingface-cli download --local-dir PersonaPlex-7B-MLX-4bit aufklarer/PersonaPlex-7B-MLX-4bit
  • Moshi

    How to use aufklarer/PersonaPlex-7B-MLX-4bit with Moshi:

    # pip install moshi
    # Run the interactive web server
    python -m moshi.server --hf-repo "aufklarer/PersonaPlex-7B-MLX-4bit"
    # Then open https://localhost:8998 in your browser
    # pip install moshi
    import torch
    from moshi.models import loaders
    
    # Load checkpoint info from HuggingFace
    checkpoint = loaders.CheckpointInfo.from_hf_repo("aufklarer/PersonaPlex-7B-MLX-4bit")
    
    # Load the Mimi audio codec
    mimi = checkpoint.get_mimi(device="cuda")
    mimi.set_num_codebooks(8)
    
    # Encode audio (24kHz, mono)
    wav = torch.randn(1, 1, 24000 * 10)  # [batch, channels, samples]
    with torch.no_grad():
        codes = mimi.encode(wav.cuda())
        decoded = mimi.decode(codes)
  • Notebooks
  • Google Colab
  • Kaggle
  • Local Apps
  • LM Studio
PersonaPlex-7B-MLX-4bit
5.82 GB
Ctrl+K
Ctrl+K
  • 1 contributor
History: 13 commits
aufklarer's picture
aufklarer
Upload README.md with huggingface_hub
8c68634 verified about 1 month ago
  • voices
    Upload PersonaPlex 7B MLX 4-bit weights 3 months ago
  • .gitattributes
    1.52 kB
    initial commit 3 months ago
  • README.md
    4.3 kB
    Upload README.md with huggingface_hub about 1 month ago
  • config.json
    1.46 kB
    Quantize depformer to int4 (2.4GB -> 698MB), change default prompt to assistant 3 months ago
  • depformer.safetensors
    731 MB
    xet
    Quantize depformer to int4 (2.4GB -> 698MB), change default prompt to assistant 3 months ago
  • embeddings.safetensors
    988 MB
    xet
    Upload PersonaPlex 7B MLX 4-bit weights 3 months ago
  • mimi.safetensors
    385 MB
    xet
    Upload PersonaPlex 7B MLX 4-bit weights 3 months ago
  • temporal.safetensors
    3.7 GB
    xet
    Quantize in_proj to int4 (3.4 GB, was 6 GB float16) 3 months ago
  • tokenizer_spm_32k_3.model
    553 kB
    xet
    Upload PersonaPlex 7B MLX 4-bit weights 3 months ago