Skip to main content

5 posts tagged with "training"

View All Tags

Running ML Experiments with Runpod and Transformer Lab

Β· 5 min read

For many looking to experiment with machine learning, the biggest barrier to entry is access to hardware. GPUs are expensive, hard to find, and even harder to share across a team. Big cloud hosting providers have complex interfaces, pricing models, and try to lock you into their ecosystem and tooling.

Runpod Logo

Using Runpod with Transformer Lab changes that. Now you can spin up GPU-backed experiments quickly from the comfort of your own system.

Transformer Lab Goes Beyond Images: Introducing Text Diffusion Model Support

Β· 4 min read

πŸŽ‰ Transformer Lab just expanded beyond image diffusion! We're thrilled to announce text diffusion model support so you can train, evaluate, and interact with cutting-edge text diffusion architectures like BERT, Dream, and LLaDA directly in Transformer Lab.

What's included in this release​

  • πŸš€ Text Diffusion Server for interactive generation with BERT, Dream, and LLaDA models
  • πŸ‹οΈ Text Diffusion Trainer for fine-tuning with masked-language and diffusion-style alignment workflows
  • πŸ“Š Text Diffusion Evaluator for benchmarking with the EleutherAI LM Evaluation Harness

Fine Tuning a Python Code Completion Model

Β· 7 min read
Person

This post details our journey to fine-tune smolLM 135M, a compact language model, for Python code completion.

We chose smolLM 135M for its size, which allows for rapid iteration. Instead of full fine-tuning, we employed LoRA (Low-Rank Adaptation), a technique that introduces trainable "adapter" matrices into the transformer layers. This provides a good balance between parameter efficiency and achieving solid results on the downstream task (code completion).

Transformer Lab handled the training, evaluation, and inference, abstracting away much of the underlying complexity. We used the flytech/python codes-25k dataset, consisting of 25,000 Python code snippets, without any specific pre-processing. Our training setup involved a constant learning rate, a batch size of 4, and an NVIDIA RTX 4060 GPU.

The Iterative Fine-tuning Process: Nine Runs to Success​

The core of this project was an iterative refinement of LoRA hyperparameters and training duration. We tracked both the training loss and conducted qualitative assessments of the generated code (our "vibe check") to judge its syntactic correctness and logical coherence. This combination of quantitative and qualitative feedback proved crucial in guiding our parameter adjustments.