Solutions to the labs and exercises in ISL.
-
Updated
Jan 21, 2019 - Jupyter Notebook
Solutions to the labs and exercises in ISL.
Introduction to Statistical Learning with Python
Using the decision tree technique based on entropy calculation, this application calculates the hit rate of the HASTIE file with a hit rate higher than 99%
HASTIE_NAIVEBAYES: from the Hastie_10_2.csv file obtained by the procedure described in https://scikit-learn.org/stable/modules/generated/sklearn.datasets.make_hastie_10_2.html, obtains a success rate in the training of 88% and 84% in the test. The main difference is that in the statistical process, each field is sampled differently according to…
Using the Sklearn classifiers: Naive Bayes, Random Forest, Adaboost, Gradient Boost, Logistic Regression and Decision Tree good success rates are observed in a very simple manner. In this work sensitivity is also considered. Treating each record individually, differences are found in the results for each record depending on the model used, which…
Taking into account that the accuracy of statistical results depend on the accuracy of the input data, not only on the algorithm, a Hastie file has been created in which all the records have the correct class assigned and tests of hit rates and sensitivity have been carried out
Introduction to Statistical Learning by Hastie, Tibshirani James, and Witten chapters' summary and lab solutions using Python3.
Add a description, image, and links to the hastie topic page so that developers can more easily learn about it.
To associate your repository with the hastie topic, visit your repo's landing page and select "manage topics."