Writing on software development, SaaS building, and the design industry.

All of my long-form thoughts on programming, tech, design, and more, collected in chronological order.

LoRA and QLoRA: Simple Fine-Tuning Techniques Explained

Fine-tuning large language models (LLMs) can be resource-intensive, requiring immense computational power. LoRA (Low-Rank Adaptation) and QLoRA (Quantized Low-Rank Adaptation) offer efficient alternatives for training these models while using fewer resources.

Customizing LLMs through Supervised Fine-tuning

In the ever-evolving domain of Natural Language Processing (NLP), supervised fine-tuning has emerged as a game-changing technique for adapting pre-trained Large Language Models (LLMs) to specific tasks. While pre-trained LLMs like those in the GPT family have made significant strides in language comprehension and generation, they often lack optimization for particular applications.

Explaining LLM Model Weights and Paramaters like I'm 10 (LLaMA)

LLama is undoubtably one of the best open-source LLM out there by Meta. If you have ever wandered how it works, how these LLMs are able to produce human like text and talk like humans. Well, you are not alone.

Install TensorFlow on your Mac M1/M2/M3 with GPU Support

Recently moved from an Intel based processor to an M1 apple silicon Mac and had a hard time setting up my development environments and tools, especially for my machine learning projects, I was particularly exited to use the new Apple Silicon ARM64 architecture and benefit from the GPU acceleration it offers for my ML tasks.