An important task to handle dataset with more number of features/dimensions.

Data keeps on increasing every second and it has become crucial to interpreting insights from this data to solve problems. And, as features of data increases so dimensions of the dataset increases. Eventually, a Machine Learning model needs to handle the complex data resulting in more complexities. On the other hand, there are a lot of features that are futile for the model or are correlated with others. Principal Component Analysis (PCA) is the way out to reduce dimensions and deduct correlated features from the dataset.

**The article is…**

Data Visualization plays a crucial role in real-time Machine Learning applications. Visualizing data makes a much easier and convenient way to know, interpret, and classify data in many cases. And there are some techniques which can help to visualize data and reduce dimensions of the dataset.

In my previous article, I gave an overview of Principal Component Analysis (PCA) and explained how to implement it. PCA is a basic technique to reduce dimensions and plot data. There are some limitations of using PCA from which the major is, it does not group similar classes together rather it is just a…

In supervised learning algorithms, labels are given to data-points (i.e- pairing a data-points with its class). In other words, models are supervised with class labels. While, in unsupervised learning algorithms, labels are not given to data-points, the model tries to find the class labels by comparing a data-point with other data-points and label similar data-points as the same class and likewise. In other words, models are not supervised explicitly.

Support Vector Machine (SVM) is a popular supervised Machine Learning algorithm used for classification problems, regression problems, and outlier detection. In simple words, when all data-points plotted in n-dimensional space (dimensions…

Machine Learning and Deep Learning models have become a credible guide to many businesses, and for a good reason. These models can guide in “predicting the future circumstances” as there are numerous methods available, and any industry can fit as per the challenges and goals one has. When we talk about Machine Learning or Deep Learning models, we are either talking about Classification (discrete output) or Regression (continuous output) problems.

- Introduction
- Accuracy
- Confusion Matrix
- F1-score
- Receiver Operating Characteristics — ROC Curve & AUC metric
- Log-loss
- R-squared / Coefficient of Determination
- Mean Absolute Percentage Error (MAPE)

While data preparation and training…

Simple explanation of keys with an example that is frequently asked in interviews and is a part of DBMS beginners tutorial.

So far, I have come across numerous articles that explain the types of Relational Database keys, and sometimes articles contain some key’s name that I haven’t even heard the name. And, I always felt confused while preparing for interviews or when someone asks questions about keys. Thus, I have prepared this article as a source with a simple explanation of each key with an example.

First, I will explain what are keys in DBMS?

**KEYS in DBMS** is an…

I have started a series to explain Exploratory Data Analysis (EDA) with a particular dataset to help to understand EDA in a better way. EDA is a broad approach & it includes different ways of implementation, it varies from dataset to dataset. To know the basics of EDA, check this article as it gives an overview of EDA while this article is written to focus on the hands-on practice of EDA rather than the basics of EDA.

- How to ensure you are ready to use machine learning algorithms in a project?
- How to choose the most suitable algorithms for your…

The goal is to turn data into information, and information into insight — Carly Fiorina

To solve any Data Science problem, it is necessary to understand raw data & somehow to convert the raw data to information for further work. Exploratory Data Analysis (EDA) is the step in which data is explored after the process of data collection. If you’re a beginner, then first be familiar with the EDA terminologies as in this article I would be covering a library namely **D-Tale** (can be used to make your work faster), which is used with just one line of code to…

“Data will talk, if you are willing to listen”- Jim Bergeson

With the proper use of data, one can gain insights and use it for numerous purposes. Raw data has no story to tell. So, to understand and gain insights from data, after the data collection process, exploratory data analysis comes into the picture. It is a crucial process to recognize patterns and understand data to prepare the model.

This article is divided into the following sections:

- Overview of Data Exploratory Analysis (EDA)
- EDA for Haberman’s dataset

The process to explore and understand data to gain insights from the data…

I explore, learn & implement concepts in Data Science. I write to understand & make complex concepts simple. https://www.linkedin.com/in/rajviishah/