--- license: apache-2.0 tags: - audio - speech - language-model - auristream library_name: transformers --- # AuriStream100M_1Pred_BigAudioDataset_500k **AuriStream** is a speech language model by **Greta Tuckute** and **Klemen Kotar**. This model predicts cochlear tokens from a tokenizer such as [WavCochCausalV8192](https://huggingface.co/TuKoResearch/WavCochCausalV8192). ## Model Details | Parameter | Value | |-----------|-------| | Parameters | ~0.09B | | Layers | 12 | | Hidden Size | 768 | | Attention Heads | 12 | | Vocab Size | 8192 | | Prediction Steps | 1 | ## Usage ```python from transformers import AutoModel, AutoConfig # Load with trust_remote_code for custom model model = AutoModel.from_pretrained( "TuKoResearch/AuriStream100M_1Pred_BigAudioDataset_500k", trust_remote_code=True, ) # Or load config first config = AutoConfig.from_pretrained("TuKoResearch/AuriStream100M_1Pred_BigAudioDataset_500k", trust_remote_code=True) ``` ## Base Model Code This checkpoint uses shared model code from [TuKoResearch/AuriStream-base](https://huggingface.co/TuKoResearch/AuriStream-base). ## Tokenizer This model uses cochlear tokens from [WavCochCausalV8192](https://huggingface.co/TuKoResearch/WavCochCausalV8192).