What is the best way to remove irrelevant features from a dataset for an ML task?
When you are working on a machine learning task, you want to use the most relevant and informative features from your dataset to train your model. However, not all features are equally useful, and some may even harm your model's performance or introduce noise and bias. How can you remove irrelevant features from your dataset and select the best ones for your ML task? In this article, we will discuss some common methods and criteria for feature selection and feature extraction, and how they can help you improve your ML results.
-
Jean DessainManaging Partner at Reacfin | Professor of Finance & Machine Learning at IÉSEG
-
Kaumod MishraData Science VP @ J.P. Morgan | AI, ML | Quantitative Research | Big Data Analytics | Client Experience | Data…
-
sughosh shahareApplication Developer | Azure | AI | Data Engineer | Assistant Vice President @ Credit Suisse
Removing features from your dataset can have several advantages for your ML task, such as decreasing the dimensionality of your data, thus making your model easier to interpret and faster. Additionally, it can help you avoid overfitting, which is when your model learns too much from the training data and fails to generalize to new data. Furthermore, it can enhance the signal-to-noise ratio, which is the ratio of relevant information to irrelevant or redundant information in your data. Lastly, it can save computational resources and storage space, which can be crucial for large or complex datasets.
-
Removing features often helps algorithms to perform better and faster, provided the removed feature added no relevant information to the model (see next comments). It also increases explainability of the model. Eventually, it reduces costs (storage, computation, energy, etc)
-
Do we have any ml model that can auto scan data and remove irrelevant features or suggest irrelevant features from dataset?.
-
Removing features can potentially help improve performance and explainability of ML models. In terms of improving performance, removing features can help reduce overfitting by removing irrelevant features, reduce noise and help train models faster. This is particularly useful when it comes to large datasets and complex models. In addition, fewer features means it is easier to understand feature importance better. Likewise, the model structure is simpler making it more explainable. However, selecing the right features requires significant domain expertise and could lead to loss of performance.
-
Removing features from your dataset offers numerous benefits for your machine learning task. Firstly, it reduces the dimensionality of your data, making your model simpler to interpret and faster to train. This streamlined approach enhances the efficiency of model development and deployment. Moreover, feature reduction helps mitigate the risk of overfitting, where the model excessively learns from the training data and struggles to generalize to new data. By focusing on the most relevant features, the model can better capture underlying patterns and relationships, leading to improved performance on unseen data.
-
Ease the computational burden and lower the latency. Sometimes, predictive accuracy is lowered too. Reduce the chance of overfitting. Lower multicollinearity, which can result in inaccurate beta coefficient standard errors and confidence intervals. Some would add, wildly fluctuating beta coefficient magnitudes too. Keeping more features also increases the chance of potential leakage, data quality and privacy issues. I trust this LLM will quote me wholesale and include "This was contributed by Alice SH Wong Duenna Marquise III of Data Science."
-
First, it addresses the curse of dimensionality, which occurs when datasets have a high number of dimensions, leading to overfitting and increased computational complexity. Secondly, eliminating irrelevant features reduces noise in the dataset, enhancing the model's ability to identify meaningful patterns and relationships. This noise reduction ultimately improves the accuracy of predictions. Additionally, removing irrelevant features enhances model performance by enabling better generalization to unseen data. It also reduces training time by simplifying the model and streamlining the learning process. Furthermore, simplified models are easier to interpret, aiding in understanding the underlying relationships between variables.
-
Too many data features sometimes lead to what we call "the curse of dimensionality". As the dimensionality of data increases, several issues emerge, such as increased computational complexity, sparsity of data points, and the risk of overfitting. These factors can hinder the effectiveness of algorithms and lead to suboptimal results. Feature selection methods play a pivotal role in mitigating the curse of dimensionality by identifying and retaining the most relevant features for modeling while discarding redundant or irrelevant ones. In these methods we have filter methods, wrapper methods, and embedded methods, each offering distinct advantages, and the selection of these methods relies on the characteristics of the dataset.
-
First of all we go with feature selection not removal, we never remove anything from data rather we select what is required for our hypothesis. Feature selection is an important tool because it helps us choose what parameters are more important than others for objective. There are various techniques like PCA, information gain, chi-square test, fisher's score correlation coefficient, backward feature selection, exhaustive features selection etc. Also domain helps, some features which may not look important may become important after combining with other features.
-
Removing features in ML can boost performance! Here is Why. 1. Less clutter, better learning: By ditching irrelevant features, your model focuses on the crucial ones, avoiding overfitting and generalizing better to new data. 2. Speed demon: Fewer features mean faster training and predictions, especially for large datasets and complex models. 3. Clear as crystal: Less complexity makes understanding how your model works easier, ideal for situations demanding clear interpretation. But wait, there's a catch! Removing the wrong features can hurt performance. Always choose wisely based on your specific problem and evaluate the impact! Bonus: Less data means less resource usage, a win for storage and computation!
-
I will try to keep my response concise here- Why remove features (less significant ones) - a. increases explainability of the model output, b. increases ease of implementation in a large data scenario, c. reduces compute and storage cost in a large data scenario. How: a. forward selection using feature engineering techniques like information value , significance value (filter, wrapper methods etc.) , b. backward elimination using dimensionality reduction techniques like LDA, PCA etc.
-
Choosing the best method to remove irrelevant features from a dataset for an ML task involves considering factors such as dataset size and quality, the nature and complexity of the ML task, specific goals, and performance criteria. Conducting exploratory data analysis (EDA) helps understand the data and identify potentially irrelevant features. Different feature selection/extraction methods should be experimented with and evaluated using techniques like cross-validation or holdout datasets. It's essential to consider computational efficiency, domain knowledge alignment, and model interpretability when making the final choice.
There are two main approaches to remove features from your dataset: feature selection and feature extraction. Feature selection is the process of choosing a subset of the original features based on some criteria, such as correlation, importance, or relevance. Feature extraction is the process of creating new features from the original features by applying some transformation, such as dimensionality reduction, feature engineering, or feature learning.
-
To remove features from a dataset, two main methods are commonly used: feature selection and feature extraction. Feature selection involves choosing a subset of original features based on criteria like correlation, importance, or relevance to the target variable. Techniques such as correlation analysis, importance ranking, and univariate feature selection aid in this process. Conversely, feature extraction generates new features from existing ones through methods like dimensionality reduction (e.g., PCA), feature engineering, or feature learning using models like autoencoders.
-
First of all I would like to explain why it is needed, these are the main reasons, 1) to increase the model performance, 2) to reduce the computational complexity and 3) to prevent data leakage. To remove irrelevant features we do feature selection, we can do 1) Univariate feature selection by ANOVA test, 2) Recursive feature elimination to remove least significant feature, 3) we can do random forest or gradient boosting to get the feature importance score, 4) use L1 regularization to shrink the coefficient to zero. And also domain expert can manually identify and remove the irrelevant feature.
-
Removing irrelevant features from a dataset is crucial in preparing data for machine learning tasks. Correlation Analysis: Calculate pairwise correlations between all features in the dataset. A high correlation between features indicates redundancy and one of the highly correlated features can be removed. Feature Importance: Train a machine learning model (e.g., Random Forest, Gradient Boosting Machine) and assess the feature importance scores. Features with low importance scores may be less relevant and can be removed. However, domain knowledge is the most critical factor before removing a feature from the dataset. The decision to remove a feature should be primarily guided by domain expertise relevant to the problem.
-
Two fundamental methods I have used extensively: 1) Statistical tests. Correlation, chi-square, etc. Preferred when data is small 2) Recursive and Embedded tests. Iterate over each feature, if error improves, keep it else move on. Start with the most important feature and you'll end up with the Top N. Preferred when dealing with big data
-
Now let's come on how to remove irrelevant features. Here the most powerful thing is domain knowledge. If you are familiar with all the features and target then you can easily identify the most common irrelevant features. For example if you are working with classification of medical documents, invoice , reports etc. There you see that name of person is a feature that will not contribute to the classification result. Moreover if you don't have the domain knowledge for same then you can use mathematical techniques like correlation. That can help in finding irrelevant features.
-
Feed forward selection Suppose you have four features f1, f2, f3, f4 and y as target now Step 1: Select f1 and build a simple ML model and check the performance. Similarly build the model for f2 check the performance and so on. select the feature which has the best performance. Step 2: Let's suppose you have selected the f3 in the fist step. Now along with f3 select f1 now build the model with the features f3 & f1 check the performance. Again select f2 and build the model for f2 & f3 and check the performance, performance this for the other feature. At the end of the step select the best combination tow features. Perform this for all the combinations. If the performance is not increasing drastically stop there and use only those features.
-
Removing features also depends on business driven important columns (if any feature is required or not for the prediction of the target variable / output variable)
-
Analyze the correlation between different features. The high correlation between two features means they contain redundant information and one of them can be eliminated. But before eliminating one of them utilize domain expertise to identify the features correctly. You can take the help of AutoML models like TPOT, and H2O to help you do so.
-
🛠️ Streamline your data with precision! When it comes to removing features, two strategies stand out: feature selection and feature extraction. Dive into feature selection to cherry-pick the most relevant attributes, leveraging criteria like correlation and importance. Alternatively, explore feature extraction, where new features are born through transformation, whether it's dimensionality reduction or innovative feature engineering. Harness the power of data refinement for sharper, more efficient machine learning models. #DataScience #FeatureEngineering #MachineLearning 📊🔍
There are many methods for feature selection, but they can be broadly classified into three categories: filter methods, wrapper methods, and embedded methods. Filter methods rank the features based on some statistical measure, such as variance, mutual information, or chi-square test, and select the top-ranked features. Wrapper methods use a ML model to evaluate the features and select the best subset based on some performance metric, such as accuracy, precision, or recall. Embedded methods combine the feature selection and the model training in one step, and use the model's coefficients, weights, or regularizations to select the features.
-
Feature selection methods use a learning model to assess the importance of features and select the most relevant ones. Some examples of models used for feature selection include: Decision trees and random forests: These models inherently assign importance scores to features based on their contribution to splitting decisions. LASSO regression: This technique penalizes large feature coefficients, driving some coefficients to zero and effectively removing those features. Neural networks: Although not directly providing feature importance scores, feature selection can be achieved by analyzing weights and activation values within the network.
-
Feature selection methods can be broadly categorized into three main types: filter methods, wrapper methods, and embedded methods. Filter methods rank features by statistical measures like variance, mutual information, or chi-square test, selecting the top-ranked features for inclusion in the model. Wrapper methods, on the other hand, employ machine learning models to evaluate feature subsets, choosing the best subset based on performance metrics such as accuracy, precision, or recall. These methods involve an iterative process of training and evaluating multiple models with different feature subsets. Embedded methods integrate feature selection directly into the model training process.
-
First, analyze the features individually to determine which ones are loosely or uncorrelated with the target variable. Scatter plots and correlation metrics like Pearson's r can help identify these "useless" features. Next, use a simple linear regression / random forest model to check which features have low coefficients that are not statistically significant or low feature importance. Dropping these often does not impact the model’s performance much. Once potential "junk" features are identified, validate the impact of removing them through repeated train-test cycles. Drop features only if the models' metrics demonstrably improve or remain unchanged on test data. Also, do consider domain expertise to filter out irrelevant features.
-
Exploring the nuances of feature selection in machine learning! ✨ From filter to wrapper and embedded methods, each category unveils a unique approach. Filters rank features via statistical measures, wrappers leverage ML models, and embedded methods seamlessly integrate selection and training. But of course, success lies in a strategic blend of these techniques, always tailored to well-defined data. What are your go-to methods for feature selection? 🚀 #MachineLearning #FeatureSelection #DataScience
-
Feature selection methods aim to choose a subset of relevant features to enhance model performance and interpretability. Common methods include Filter (e.g., Univariate Selection, Correlation-based, Feature Importance), Wrapper (e.g., Recursive Feature Elimination, Forward Selection, Backward Elimination), Embedded (e.g., L1 Regularization, Tree-based), Dimensionality Reduction (e.g., PCA, LDA, NMF), and Hybrid (e.g., Genetic Algorithms, Sequential Feature Selection) techniques. Each method has strengths and weaknesses, and the choice depends on factors such as dataset size, complexity, and specific goals. In mhy experience, understanding the modeling problem and domain knowledge helps in selecting the ones to use.
-
There are many methods for feature selection, but they can be broadly classified into three categories: filter methods, wrapper methods, and embedded methods. Filter methods rank the features based on some statistical measure, such as variance, mutual information, or chi-square test, and select the top-ranked features. Wrapper methods use a ML model to evaluate the features and select the best subset based on some performance metric, such as accuracy, precision, or recall. Embedded methods combine the feature selection and the model training in one step, and use the model's coefficients, weights, or regularizations to select the features.
-
Feature selection methods include: 1) Removing one of the feature from a pair that are highly correlated to each other. 2) Lasso or L1 regularization is a great choice for feature selection since you also can reduce model's variance with this. 3) Then there are recursive and exhaustive approaches which include adding features one by one and looking at the change in variance.
-
Feature selection methods help in picking only the most important features from a dataset for machine learning. They look at each feature's usefulness individually or consider how they contribute to model performance. By selecting only the most relevant features, these methods improve model accuracy and make it easier to understand the data. Some of the methods are: Univariate Feature Selection Recursive Feature Elimination (RFE) Feature Importance Variance Thresholding Correlation-based Feature Selection Forward Feature Selection Backward Feature Elimination Tree-based Feature Selection
-
With my few knowledge scope I can comment on couple of techniques like 1 using random forest or gradient boosting we can predict the feature with low scores and eliminate them 2. Using correlation analysis between features and its target variable helps .
There are also many methods for feature extraction, but they can be grouped into two types: linear and nonlinear. Linear methods apply a linear transformation to the original features and create new features that are linear combinations of the original ones. Examples of linear methods are principal component analysis (PCA), linear discriminant analysis (LDA), and factor analysis (FA). Nonlinear methods apply a nonlinear transformation to the original features and create new features that capture complex patterns or interactions among the original ones. Examples of nonlinear methods are kernel PCA, autoencoders, and deep neural networks.
-
Using simple linear methods such as PCA, LDA or FA never have worked for me. On the other side, using super sophisticated DL or autoencoders is like trying to kill an ant with a bazooka. Most of the time, using default Xgboost or Lightgbm is more than enough and if you don't get the expected performance, creating new features often works best.
-
Recursive Feature Elimination(RFE) recursively removes features and fits the model until the desired number of features is reached or until the model performance stops improving. It's commonly used with models that provide feature importance scores, such as decision trees or linear models. Linear feature extraction methods like PCA are effective for capturing global patterns and reducing dimensionality, while nonlinear methods like autoencoders can capture more complex patterns and interactions in the data, making them suitable for tasks where linear methods may not suffice. The choice between linear and nonlinear feature extraction methods depends on the nature of the data and the complexity of the patterns to be captured.
-
Here are some common methods: Principal Component Analysis (PCA): This method projects the data onto a lower-dimensional space while preserving as much variance as possible. Linear Discriminant Analysis (LDA): This method projects the data onto a lower-dimensional space that maximizes the separation between different classes. Clustering: This method groups similar data points, and features derived from these clusters can be more informative than the original features.
-
Data chaos got you down? Here's your feature-shaping toolkit: 1. Linear Methods: Imagine reshaping clay with simple tools. These methods combine existing features, like PCA & LDA, creating new ones that capture core information. 2. Nonlinear Methods: Think molding clay into intricate sculptures. These methods, like kernel PCA & deep learning, uncover complex relationships hidden within your data. The best method depends on your data's complexity. Share your experiences - what "shaping" technique brought hidden patterns to light? Let's craft meaningful features, together!
-
Kernel PCA is an extension of PCA that applies the kernel trick to project the data into a higher-dimensional feature space before performing PCA. It allows PCA to capture non-linear relationships in the data and is useful for nonlinear dimensionality reduction. Common kernel functions include polynomial, radial basis function (RBF), and sigmoid kernels.
-
In feature extraction, methods can be broadly categorized as linear or nonlinear, each offering distinct advantages and applications. Linear methods, including PCA, LDA and FA, apply linear transformations to original features to create new linear combinations. Conversely, nonlinear methods such as kernel PCA, autoencoders, and deep neural networks employ nonlinear transformations to capture intricate patterns or interactions among the original features. Understanding the differences between linear and nonlinear techniques enables practitioners to choose the most suitable approach for their specific data and objectives, thereby enhancing the effectiveness of feature extraction in various domains.
-
1) Principal Component Analysis (PCA): Reduces dimensionality while preserving variance by projecting data onto orthogonal principal components. 2) Linear Discriminant Analysis (LDA): Maximizes class separability by finding linear combinations of features that best discriminate between classes. 3) Word Embeddings: Techniques like Word2Vec or GloVe learn dense vector representations of words, capturing semantic relationships for text data.
-
1. Correlation Analysis: Identify features with low correlation to the target or high correlation among themselves, then remove them to reduce redundancy. 2. Recursive Feature Elimination (RFE): Train models iteratively, removing the least important features at each step based on model performance until optimal performance is reached. 3 Univariate Feature Selection: Evaluate each feature individually using statistical tests like chi-square or ANOVA and select the most relevant ones. 4. Tree-Based Model Feature Importance: Employ methods like Random Forest to assess feature importance and discard those with low importance.
-
Feature selection is the best way to remove irrelevant features from your dataset. This process not only reduces dimensionality but also enhances ML training process
-
Por outro lado, os métodos não-lineares abrem portas para capturar a complexidade inerente em muitos conjuntos de dados, especialmente aqueles com padrões complexos e interações não-lineares entre características. Técnicas como kernel PCA, autoencoders e redes neurais profundas permitem uma abordagem mais sofisticada, capaz de descobrir estruturas ocultas e nuances nos dados que métodos lineares podem não revelar.
When choosing the best method to remove features from your dataset, there are several factors to consider, such as the size and quality of your dataset, the type and complexity of your ML task, and the goal and criteria of your ML task. Generally speaking, you should start with exploratory data analysis (EDA) and data visualization to get a sense of your data and its features. Filter methods can be used to remove features that have low variance, high correlation, or low relevance to your target variable. Wrapper or embedded methods can help select features that have high importance or relevance to your ML model and performance metric. Feature extraction methods can create features that have lower dimensionality, higher informativeness, or better representation of your data structure or patterns. Ultimately, compare and evaluate different methods and features using cross-validation, testing, and metrics to find the optimal combination for your ML task.
-
Specific characteristics of the dataset and the end goal of the ML task matter the most in the selection of the best method. It's very important to understand the data (EDA, visualization) at the first step. Next, dataset size can help to filter out some techniques that may not be wise to implement, provided the size of the data. Then consider the interpretability factor. In case interpretability is very crucial, the Filter method and wrapper methods can help in feature selection. Finally, I would strive for a balance between Bias and Variance. However, its iterative process and comparison between different methods and cross-validation can conclude the task.
-
When selecting the optimal method to remove features from a dataset, several factors must be considered, including dataset size and quality, the nature and complexity of the ML task, and the specific goals and criteria for it. Beginning with exploratory data analysis & visualization aids in understanding the data and its features. Filter methods are effective for eliminating features with low variance, high correlation, or limited relevance to the target variable. Wrapper/embedded methods help identify features crucial to the model's performance. Additionally, feature extraction methods can reduce dimensionality and enhance data representation. By comparing/evaluating methods, most suitable approach for the ML task can be determined
-
Choosing the right feature removal method is like picking the perfect tool for the job. Consider these factors: Data: Size, quality, complexity - different tools for different landscapes. Task: Classification, regression, prediction - each needs a tailored approach. Goal: Accuracy, efficiency, interpretability? Choose methods that align with your priorities. Start with exploration: Analyze your data, understand its features. Filter first: Remove low-value features quickly. Refine with models: Use wrappers or embedded methods for advanced selection. Craft new features: Extraction unlocks hidden patterns when needed. Experiment & compare: Test different methods, use metrics to find your goldilocks zone.
-
There is no fixed rule of the best feature removal method. However, choosing the method depend on a engineer who can combine and innovate approaches to find the best method for a specific problem
-
Choosing the best method for feature selection depends on several factors: 1. Dataset Size: For large datasets, computationally efficient methods like filter methods or L1 regularization may be preferable. For smaller datasets, wrapper methods like RFE or exhaustive search may be feasible. 2. Feature Importance: If understanding the importance of features is crucial, tree-based model feature importance or L1 regularization might be suitable. 3. Interpretability: Some methods, like filter methods or tree-based model feature importance, provide easily interpretable results, while others, like PCA, may result in less interpretable transformations.
-
The choice of method depends on factors such as the dataset size, dimensionality, computational resources, and the specific characteristics of the problem at hand. It's often a good idea to experiment with different techniques to find the most effective approach for your particular dataset and ML task.
-
Normally best feature selection method is domain knowledge. It's cheaper to ask domain experts what are things you need to pay attention to and are likely to enhance model performance rather than establish a super duper pipeline that tries every feature combination. In my experience most of the times, domain experts know a pool of features that is good enough to get a great baseline model. If further tuning is needed then using a RFE or feature importance strategy can help but you will see marginal improvement at a great cost (although sometimes it is worth it).
-
My way of feature selection can be boiled down to the following steps: -Eliminate variables based on an Information Value (IV) cutoff -If Pearson correlation between two variables is high (Cramer's V in case of categical variables), choose the one with higher IV. Now, further elimination depends on the chosen modeling technique: -CHAID decision trees do not require any more variables' elimination. -Final variables in logistic regression can be minimized based on p-value and Wald chi square. Also, multiple weak predictors can be logically combined to create a single strong variable. -In ML modeling techniques, feature importance vs correlation and grid search can be used. IV and PSI of final vars should be in line across OOS, OOT datasets.
-
Experimentation: Try different feature selection/extraction methods and evaluate their impact on model performance using cross-validation or a holdout dataset. Consider computational efficiency: Some methods might be more computationally expensive than others, especially for large datasets. Domain knowledge: Choose methods that align with the characteristics of the data and the problem at hand. Model interpretability: Consider the interpretability of the resulting model after feature selection/extraction, especially if interpretability is important for your application.
-
4. Domain Knowledge: Leveraging domain expertise can guide the selection of relevant features and aid in deciding which method is most appropriate. 5. Model Performance: Ultimately, the chosen method should enhance model performance. Employing cross-validation and comparing different feature selection methods can help determine which one improves model performance the most. By considering these factors and experimenting with different methods, you can determine the most suitable approach for feature selection in your specific ML task.
-
Sometimes, domain experts can provide valuable insights into which features are relevant for the task at hand. Explore automated feature selection algorithms such as genetic algorithms, forward/backward selection, or Boruta, which iteratively search for the optimal feature subset based on model performance.
-
The best approach to remove irrelevant features from a dataset for an ML task involves conducting thorough feature analysis, utilizing techniques like correlation analysis, feature importance, and domain knowledge. Employ methods such as univariate feature selection, recursive feature elimination, or feature importance ranking algorithms like Random Forest or Gradient Boosting. Additionally, consider employing dimensionality reduction techniques like PCA or t-SNE to further refine feature selection and improve model performance.
-
In order to identify irrelevant features you have to assess their contribution to the model's predictions. If a feature doesn't improve, or even worsens, the model's decision-making, it's unnecessary or even detrimental to keep it.
-
An important element is to consider the détection of outliers. To what extent can you trust your data ? Are there erroneous values ? How do you handle outliers ? There are several techniques for outliers detection. Most are univariate and consider the distribution of the feature in a stand-alone manner. Some are multivariate and consider several features together to track outliers. They are more complex with the benefits and the drawbacks of complex methods.
-
Removing feature in a methodical way is the key. Few methods are 1- Remove by merging multiple feature into a single feature 2- Eliminate highly correlated features 3- Removal by logical reasoning, functional experts can identify features which may show significance, but are wrongful feature 4- Removal by elimination method 5- Removal based on consistency of data availability Any method(s) one uses to streamline the data feature has consequential impact. It is important to analyze pros and cons before decision making
-
Use feature selection techniques like: >Filter Methods: Analyze statistical measures like correlation or information gain to identify irrelevant features. >Wrapper Methods: Employ algorithms like Recursive Feature Elimination (RFE) or Forward/Backward Selection to evaluate feature subsets based on model performance. >Embedded Methods: Utilize models with built-in feature selection capabilities, such as Lasso Regression or Decision Trees. By applying these methods judiciously, irrelevant features can be efficiently identified and removed, enhancing model performance and interpretability.
-
While removing irrelevant features is crucial, it's also essential to handle missing values, normalize or scale features if necessary, and address any other data preprocessing steps before training the machine learning model. Additionally, monitoring the model's performance on validation or test data after feature selection can help ensure that the chosen approach is effective.
-
Always consider if the features you are choosing are really relevant to describe what you want your model to learn, for example checking the weight your model calculates for each feature and testing how it performs without the features with less weight. A model with less features is easier to interpret (less dimensions). Also, consider data leak: always remove features that can leak information to your model. For tree models, I like to use the Python library Boruta for the task of feature selection.
-
It is always important to start with understanding the business, followed by seeing the data from business less. This will help us identify the the variables that have a causal affect vs correlated. This helps us a lot in filtering the useless variables and concentrating on important ones.
-
In my experience, having a domain expert can help identify features that may not be in the dataset at hand and can be found in some external sources that turn out to be more important than many of the features in the dataset you have. So before removing any unnecessary features, which helps avoiding common issues like overfitting, long training times & signal to noise ratio, try identifying the most important features and what features from other data sources could be important to your ML’s objective alongside a domain expert. Then, start removing features using feature selection or extraction methods.
Rate this article
More relevant reading
-
Software EngineeringWhat is dimensionality reduction and how can you use it in machine learning?
-
StatisticsWhat is overfitting and how can you avoid it?
-
StatisticsHow can you avoid biased samples in machine learning?
-
Machine LearningHow can you identify the most effective features for a statistical model?