This guide dives into setting up large-scale machine learning experiments with a clean code structure, local testing, and efficient parallelization using tools like Typer, Pydantic, and Argo Workflows. Learn how to move beyond notebooks, structure ML projects for scalability, and log results effectively to accelerate model development.
Optimization tricks used for speeding up transformer inference
Tutorial on finetuning LLMs via HF transformers library with wandb logging
This study finds NuExtract performs best for structured outputs, with KV caching improving speed and accuracy for larger models despite some hallucinations.
Using Small Language Models to with Small vision models to generate captions
MLOps: Leveraging ArgoWF and Buildkite to train models
Using KV caching and logit ratios to speed up and control LLM/ VLM outputs.
Training MNIST via DDPM
Fine-tuning T5 for Sequence to Sequence tasks
Fine-tuning GPT2 for Sequence to Sequence tasks
A failed attempt at model compression using student teacher learning
Python testing for Machine Learning
Implementing Approximate Nearest Neighbours Oh Yeah (ANNOY)
Implementing kmeans with cosine distance
How prefetch_factor did not help in streaming data
Using Encoder Decoder models in HF to combine vision and text
Timing comparison of tokenizer as collate function and after batching
Fast Zero Shot classification of text
Getting image embeddings with no negative samples
Getting image patches for Visual Transformer
PyTorch Collate function tutorial
Creating TabNet from Scratch in Tensorflow 2.0
Training OpenAI’s CLIP on google colab
Using Callbacks to get Optimal Learning Rate
Extending normal Focal Loss