scikit-learn Tutorials

Dive into the world of machine learning with LabEx's curated scikit-learn tutorials. Discover practical applications, explore cutting-edge techniques, and hone your skills through interactive coding exercises.

Pairwise Metrics and Kernels in Scikit-Learn
Pairwise Metrics and Kernels in Scikit-Learn
In this lab, we will explore the sklearn.metrics.pairwise submodule in scikit-learn. This module provides utilities for calculating pairwise distances and affinities between sets of samples.
Machine Learningscikit-learn
Kernel Approximation Techniques in Scikit-Learn
Kernel Approximation Techniques in Scikit-Learn
This tutorial will guide you through the process of using kernel approximation techniques in scikit-learn.
Machine Learningscikit-learn
Transforming the Prediction Target
Transforming the Prediction Target
In machine learning, it is often necessary to transform the prediction target before training a model. This can include tasks such as converting multiclass labels into a binary indicator matrix or encoding non-numerical labels into numerical labels.
Machine Learningscikit-learn
Partial Dependence and Individual Conditional Expectation
Partial Dependence and Individual Conditional Expectation
Partial dependence plots (PDP) and individual conditional expectation (ICE) plots are useful tools for visualizing and analyzing the interaction between the target response and a set of input features. PDPs show the dependence between the target response and the input features, while ICE plots visualize the dependence of the prediction on a feature for each individual sample. These plots help us understand the relationship between the target response and the input features.
Machine Learningscikit-learn
Permutation Feature Importance
Permutation Feature Importance
In this lab, we will learn about the Permutation Feature Importance method, which is a model inspection technique used to determine the importance of features in a predictive model. This technique can be especially useful for non-linear or opaque models that are difficult to interpret.
Machine Learningscikit-learn
Imputation of Missing Values
Imputation of Missing Values
Many real-world datasets contain missing values, which can cause issues when using machine learning algorithms that assume complete and numerical data. In such cases, it is important to handle missing values appropriately to make the most of the available data. One common strategy is imputation, which involves filling in the missing values based on the known part of the data.
Machine Learningscikit-learn
Preprocessing Techniques in Scikit-Learn
Preprocessing Techniques in Scikit-Learn
In this lab, we will explore the preprocessing techniques available in scikit-learn. Preprocessing is an essential step in any machine learning workflow as it helps to transform raw data into a suitable format for the learning algorithm. We will cover various preprocessing techniques such as standardization, scaling, normalization, encoding categorical features, imputing missing values, generating polynomial features, and creating custom transformers.
Machine Learningscikit-learn
Validation Curves: Plotting Scores to Evaluate Models
Validation Curves: Plotting Scores to Evaluate Models
In machine learning, every estimator has its advantages and drawbacks. The generalization error of an estimator can be decomposed into bias, variance, and noise. The bias of an estimator is the average error for different training sets, while the variance indicates its sensitivity to varying training sets. Noise is a property of the data.
Machine Learningscikit-learn
Pipelines and Composite Estimators
Pipelines and Composite Estimators
In scikit-learn, pipelines and composite estimators are used to combine multiple transformers and estimators into a single model. This is useful when there is a fixed sequence of steps for processing the data, such as feature selection, normalization, and classification. Pipelines can also be used for joint parameter selection and to ensure that statistics from the test data do not leak into the trained model during cross-validation.
Machine Learningscikit-learn
Feature Extraction with Scikit-Learn
Feature Extraction with Scikit-Learn
In this lab, we will learn how to perform feature extraction using the scikit-learn library. Feature extraction is the process of transforming raw data into numerical features that can be used by machine learning algorithms. It involves extracting relevant information from different types of data such as text and images.
Machine Learningscikit-learn
Decomposing Signals in Components
Decomposing Signals in Components
In this lab, we will explore the topic of decomposing signals into components using matrix factorization techniques provided by scikit-learn. We will cover techniques such as Principal Component Analysis (PCA), Independent Component Analysis (ICA), Non-negative Matrix Factorization (NMF), and more. This lab will guide you through the process of decomposing signals into their components step by step.
Machine Learningscikit-learn
Density Estimation Using Kernel Density
Density Estimation Using Kernel Density
In this lab, we will explore density estimation, which is a technique used to estimate the probability density function of a random variable. Specifically, we will focus on kernel density estimation, which is a non-parametric method for estimating the density.
Machine Learningscikit-learn
Evaluating Machine Learning Model Quality
Evaluating Machine Learning Model Quality
In machine learning, it is important to evaluate the quality of the predictions made by a model. This helps us understand how well the model is performing and whether it can be trusted for making accurate predictions. The scikit-learn library provides several metrics and scoring methods to quantify the quality of predictions.
Machine Learningscikit-learn
Biclustering in Scikit-Learn
Biclustering in Scikit-Learn
Biclustering is a method that simultaneously clusters rows and columns of a data matrix. This allows us to identify submatrices within the data matrix that have specific properties. Biclustering is useful in various domains, including data analysis, image processing, and bioinformatics.
Machine Learningscikit-learn
Tuning Hyperparameters of an Estimator
Tuning Hyperparameters of an Estimator
Hyperparameters are parameters that are not directly learned by an estimator. They are passed as arguments to the constructor of the estimator classes. Tuning the hyperparameters of an estimator is an important step in building effective machine learning models. It involves finding the optimal combination of hyperparameters that result in the best performance of the model.
Machine Learningscikit-learn
Machine Learning Cross-Validation with Python
Machine Learning Cross-Validation with Python
In machine learning, cross-validation is a technique used to evaluate the performance of a model on an independent dataset. It helps to prevent overfitting by providing a better estimate of how well the model will generalize to new, unseen data.
Machine Learningscikit-learn
Covariance Matrix Estimation with Scikit-Learn
Covariance Matrix Estimation with Scikit-Learn
Covariance estimation is an important statistical technique used to estimate the covariance matrix of a population. The covariance matrix describes the relationship between variables in a dataset and can provide valuable insights into the data's scatter plot shape. In this lab, we will explore various methods for estimating the covariance matrix using the sklearn.covariance package in Python.
Machine Learningscikit-learn
Neural Network Models
Neural Network Models
In this lab, we will learn about neural network models and how they can be used in supervised learning tasks. Neural networks are a popular type of machine learning algorithm that can learn non-linear patterns in data. They are often used for classification and regression tasks.
Machine Learningscikit-learn
  • Prev
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • ...
  • 15
  • Next