Feature Engineering

Feature Engineering

Professional Development

10 Qs

quiz-placeholder

Similar activities

SQL basics

SQL basics

Professional Development

12 Qs

Intro SQL 2 - Data Dictionary

Intro SQL 2 - Data Dictionary

Professional Development

12 Qs

Power BI Mastery

Power BI Mastery

Professional Development

15 Qs

FinTech 02-3 Python

FinTech 02-3 Python

Professional Development

10 Qs

Pre-Quiz Self Assessment

Pre-Quiz Self Assessment

Professional Development

15 Qs

Qualitative Data

Qualitative Data

University - Professional Development

15 Qs

JMP Quizerable

JMP Quizerable

Professional Development

10 Qs

Daily Quiz (24.12.20)

Daily Quiz (24.12.20)

7th Grade - Professional Development

10 Qs

Feature Engineering

Feature Engineering

Assessment

Quiz

Other

Professional Development

Easy

Created by

Bayu Prasetya

Used 11+ times

FREE Resource

10 questions

Show all answers

1.

OPEN ENDED QUESTION

15 mins • 1 pt

What is feature engineering?

Evaluate responses using AI:

OFF

Answer explanation

Feature engineering is the process of selecting and transforming raw data features into new features that are more informative, useful, and predictive for machine learning models. In other words, it involves identifying and creating new features from the raw data that can improve the performance of a machine learning model.

The goal of feature engineering is to create a set of features that best represents the underlying patterns and relationships in the data, and that can improve the accuracy of a machine learning model. Feature engineering is a crucial step in machine learning, as the quality of the input features can have a significant impact on the model's ability to learn and make accurate predictions.

Some common techniques used in feature engineering include:

1. Feature scaling or normalization

Handling missing data

2. Encoding categorical variables

3. Creating new features from existing ones (e.g. combining features, binning, scaling, etc.)

4. Feature selection (e.g. removing irrelevant or redundant features)

Effective feature engineering requires a deep understanding of the data and the problem at hand, as well as knowledge of the various techniques available. It can be a time-consuming process, but it is often essential for achieving high performance in machine learning models.

2.

OPEN ENDED QUESTION

15 mins • 1 pt

Explain what scaling means, and when we use it?

Evaluate responses using AI:

OFF

Answer explanation

Scaling refers to the process of transforming the numerical values of features in a dataset to a common scale, typically between 0 and 1, or between -1 and 1. The purpose of scaling is to ensure that all the features have the same weight and importance during training a machine learning model.

When we use scaling: In many machine learning algorithms, the magnitude and distribution of the features can have a significant impact on the performance of the model. Features with larger magnitudes can dominate the model, leading to biased results. Additionally, some algorithms like K-Nearest Neighbors (KNN) and Support Vector Machines (SVM) use distance measures to determine similarity between data points, and features with larger magnitudes can heavily influence the distance calculations.

In summary, scaling is a crucial step in preparing data for machine learning models, as it ensures that all features are treated equally during training and prevents biases caused by differences in feature magnitudes.

3.

OPEN ENDED QUESTION

15 mins • 1 pt

What is an outlier, and what is a proper scaling method for data that have outlier values?

Evaluate responses using AI:

OFF

Answer explanation

Outliers are data points that are significantly different from the majority of the data points in a dataset. When scaling data that contains outliers, it is important to choose a scaling method that is robust to outliers and does not heavily influence the scaling of the rest of the data.

One such method is the Robust Scaler, which is a scaling method that is robust to outliers. The RobustScaler works by scaling the data using the interquartile range (IQR), which is the difference between the 75th and 25th percentiles of the data. The RobustScaler then scales the data such that the IQR of each feature is 1, and the median of each feature is 0.

4.

OPEN ENDED QUESTION

15 mins • 1 pt

Explain what One Hot Encoding is and what is the difference between One Hot Encoding and Binary Encoding!

Evaluate responses using AI:

OFF

Answer explanation

One hot encoding: One hot encoding is a process of representing categorical variables as a binary vector. It involves creating a new binary feature for each possible category of the original categorical feature. The value of the binary feature is 1 if the data point belongs to that category, and 0 otherwise. For example, if we have a categorical variable "color" with possible values "red," "blue," and "green," one hot encoding would create three new binary features "color_red," "color_blue," and "color_green."

inary encoding: Binary encoding is similar to one hot encoding, but it represents each category as a binary number rather than a binary vector. It involves assigning each category a unique number and then representing that number as a binary vector. For example, if we have a categorical variable "color" with possible values "red," "blue," and "green," binary encoding might assign the numbers 0, 1, and 2 to these categories and represent them as binary vectors (0 = 00, 1 = 01, 2 = 10).

The main difference between one hot encoding and binary encoding is that one hot encoding creates a new binary feature for each category, while binary encoding represents each category as a binary number. This means that binary encoding uses fewer features and may be more memory-efficient, but it may also be less expressive and have lower accuracy in some cases.

5.

OPEN ENDED QUESTION

15 mins • 1 pt

If we encode the data with 7 categories,

1. How many new columns are formed if we use One Hot Encoding?

2. How many new columns are formed if we use One Hot Encoding with the parameter drop_first = True (sklearn)?

3. How many new columns are formed if we use the Binary Encoder?

Evaluate responses using AI:

OFF

Answer explanation

1. 7 columns

2. 6 columns

3. 4 clumns

6.

OPEN ENDED QUESTION

15 mins • 1 pt

If we have a dataset that comes from survey results (eg customer satisfaction) which consists of 3 categories: Dissatisfied, Satisfied, Very Satisfied. What encoding method is suitable for the data, and explain why

Evaluate responses using AI:

OFF

Answer explanation

For categorical data like customer satisfaction levels, Ordinal encoding is a suitable encoding method. Ordinal encoding can be useful when there is a natural ordering to the categories, such as in the case of customer satisfaction levels or education levels. The categories "Dissatisfied," "Satisfied," and "Very Satisfied" have a natural ordering, and assigning numerical values to each category in a way that preserves this ordering can be useful for some machine learning algorithms.

However, it is important to note that ordinal encoding assumes that the distance between the values is equal, which may not be true in all cases. For example, the difference between Dissatisfied and Satisfied may not be the same as the difference between Satisfied and Very Satisfied.

7.

OPEN ENDED QUESTION

15 mins • 1 pt

What is feature selection, and how do we do feature selection?

Evaluate responses using AI:

OFF

Answer explanation

Feature selection is the process of selecting a subset of relevant features (i.e., variables or attributes) from a larger set of features in a dataset, with the goal of improving the performance of a machine learning model.

There are several methods for feature selection, including:

1. Filter methods: These methods use statistical measures (e.g., correlation, mutual information) to rank the importance of each feature and select the top k features. Filter methods are generally fast and computationally efficient, but they do not consider the interaction between features.

2. Wrapper methods: These methods evaluate the performance of a machine learning model using different subsets of features and select the subset that achieves the best performance. Wrapper methods can be computationally expensive, but they can capture the interaction between features.

3. Embedded methods: These methods incorporate feature selection into the process of building a machine learning model, such as using regularization techniques (e.g., Lasso, Ridge) that penalize the coefficients of less important features. Embedded methods can be computationally efficient and effective, but they may not be suitable for all types of machine learning models.

4. Domain knowledge: Sometimes, domain knowledge can be used to select relevant features based on the understanding of the problem domain and the potential impact of each feature on the outcome.

When selecting features, it is important to consider the trade-off between the number of features and the performance of the machine learning model. Having too many irrelevant features can lead to overfitting and reduce the performance of the model, while having too few relevant features can result in underfitting and also reduce the performance

Create a free account and access millions of resources

Create resources
Host any resource
Get auto-graded reports
or continue with
Microsoft
Apple
Others
By signing up, you agree to our Terms of Service & Privacy Policy
Already have an account?