Feature Selection for Machine Learning

September 29, 2021 | 8:49 am
Soledad Galli, Lead Data Scientist, Train in Data

Please complete the form to watch the webinar recording


    You can watch the video below. Let us know if you learn something new or useful by tweeting us @DataIdols.

    When we build machine learning models with the aim to use them in production, we probably don’t want to use all the variables available in the data. Sure, adding more variables rarely makes a model less accurate, but there are certain disadvantages to including an excess of features. To select the most predictive variables, we can use several feature selection algorithms. They are typically grouped in 3 categories, filter, wrapper and embedded methods, and those algorithms that do not fit in these categories are sort of hybrid methods. In this video, I first discuss the importance of feature selection and then go through the categories of feature selection methods and describe the most popular algorithms of each. I will also compare the implementation of these feature selection algorithms in open source Python libraries.

    Mmm 🍪cookies!

    We use cookies to make your experience on this website better, and we have a variety to choose from. Use the toggles below to customise your selection or click 'Save my cookies' to get straight to the content.