Llama 4 Debug Model - Trained with TorchTitan

Custom-trained Llama 4 model using TorchTitan framework.

Model Details

  • Training Framework: TorchTitan
  • Training Steps: 10,000
  • Model Size: ~220 MB
  • Precision: bfloat16

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("lakhera2023/llama4-debugmodel-10k") tokenizer = AutoTokenizer.from_pretrained("lakhera2023/llama4-debugmodel-10k")

prompt = "Once upon a time" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=100) print(tokenizer.decode(outputs[0]))

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for lakhera2023/llama4-debugmodel-10k

Finetuned
(19)
this model