Skip to main content

Posts

Featured

Understanding the Decision Tree [ Day 1]

 Hello Folks, Today we will explore the Decision Tree in Machine Learning. The decision tree is usually called CART (Classification and Regression Trees). 1.DecisionTreeClassifier: class   sklearn.tree. DecisionTreeClassifier ( * ,  criterion = 'gini' ,  splitter = 'best' ,  max_depth = None ,  min_samples_split = 2 ,  min_samples_leaf = 1 ,  min_weight_fraction_leaf = 0.0 ,  max_features = None ,  random_state = None ,  max_leaf_nodes = None ,  min_impurity_decrease = 0.0 ,  class_weight = None ,  ccp_alpha = 0.0 ) Advantages : 1.Easy to Understand  Disadvantages: Decision trees tend to overfit data with a large number of features. Getting the right ratio of samples to a number of features is important since a tree with few samples in high dimensional space is very likely to overfit. Purning: (limit the height of tree) Pruning is a data compression technique in machine learning and search algorithms that reduces the size of decision trees by removing sections of the tr

Latest posts

Static keyword in C

Binary Search And Linear search

Starting with Algorithm