Llama 4 Debug Model - Trained with TorchTitan
Custom-trained Llama 4 model using TorchTitan framework.
Model Details
- Training Framework: TorchTitan
- Training Steps: 10,000
- Model Size: ~220 MB
- Precision: bfloat16
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("lakhera2023/llama4-debugmodel-10k") tokenizer = AutoTokenizer.from_pretrained("lakhera2023/llama4-debugmodel-10k")
prompt = "Once upon a time" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=100) print(tokenizer.decode(outputs[0]))
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for lakhera2023/llama4-debugmodel-10k
Base model
meta-llama/Llama-4-Scout-17B-16E