Interpreting Coefficients in Linear Regression Models

Linear regression models are foundational in machine learning. Merely fitting a straight line and reading the coefficient tells a lot. But how do we extract and interpret the coefficients from these models to understand their impact on predicted outcomes? This post will demonstrate how one can interpret coefficients by exploring various scenarios. We’ll delve into …

Basic Statistical Analysis with NumPy

Introduction Statistical analysis is important in data science. It helps us understand data better. NumPy is a key Python library for numerical operations. It simplifies and speeds up this process. In this article, we will explore several functions for basic statistical analysis offered by NumPy. NumPy is a Python library for numerical computing. It helps …

3 Ways of Using Gemma 2 Locally

After the highly successful launch of Gemma 1, the Google team introduced an even more advanced model series called Gemma 2. This new family of Large Language Models (LLMs) includes models with 9 billion (9B) and 27 billion (27B) parameters. Gemma 2 offers higher performance and greater inference efficiency than its predecessor, with significant safety …

7 Machine Learning Projects That Can Add Value to Any Resume

Learning by doing is the best way to master essential skills for becoming a machine learning engineer. Instead of just focusing on simple classification and regression models. In this blog, we will focus on advanced machine learning projects that will impact your resume and attract recruiters and hiring managers. We will learn about computer vision …

One Hot Encoding: Understanding the “Hot” in Data

Preparing categorical data correctly is a fundamental step in machine learning, particularly when using linear models. One Hot Encoding stands out as a key technique, enabling the transformation of categorical variables into a machine-understandable format. This post tells you why you cannot use a categorical variable directly and demonstrates the use One Hot Encoding in …

The Search for the Sweet Spot in a Linear Regression with Numeric Features

Consistent with the principle of Occam’s razor, starting simple often leads to the most profound insights, especially when piecing together a predictive model. In this post, using the Ames Housing Dataset, we will first pinpoint the key features that shine on their own. Then, step by step, we’ll layer these insights, observing how their combined …

5 Free Podcasts That Demystify Machine Learning Concepts

Machine learning (ML) has become a buzzword in recent years, with applications ranging from voice assistants to self-driving cars. Yet, for many, the inner workings of these technologies remain a mystery. Podcasts offer a great way to learn about this field without getting overwhelmed. They break down complex ideas into simpler terms and let you …

Building a Simple RAG Application Using LlamaIndex

In this tutorial, we will explore Retrieval-Augmented Generation (RAG) and the LlamaIndex AI framework. We will learn how to use LlamaIndex to build a RAG-based application for Q&A over the private documents and enhance the application by incorporating a memory buffer. This will enable the LLM to generate the response using the context from both …

The Strategic Use of Sequential Feature Selector for Housing Price Predictions

To understand housing prices better, simplicity and clarity in our models are key. Our aim with this post is to demonstrate how straightforward yet powerful techniques in feature selection and engineering can lead to creating an effective, simple linear regression model. Working with the Ames dataset, we use a Sequential Feature Selector (SFS) to identify …