The curse of dimensionality refers to the potential dangers associated with modeling from a dataset that has a large number of features. The risk for overfitting increases with the number of input features, largely due to the likelihood of data sparsity occurring. It is also increasingly difficult to meaningfully interpret the relationships learned in a machine learning algorithm when the number of features is large. Feature selection and dimensionality reduction are the most appropriate remedies for making sense out of high dimensional data.
What is the Curse of Dimensionality?
Help us improve this post by suggesting in comments below:
– modifications to the text, and infographics
– video resources that offer clear explanations for this question
– code snippets and case studies relevant to this concept
– online blogs, and research publications that are a “must read” on this topic
Partner Ad