LDR | | 05139cmm u2200481Mu 4500 |
001 | | 000000316251 |
003 | | OCoLC |
005 | | 20230525180049 |
006 | | m d |
007 | | cr cnu---unuuu |
008 | | 190622s2019 xx o 000 0 eng d |
020 | |
▼a 1838820647 |
020 | |
▼a 9781838820640
▼q (electronic bk.) |
035 | |
▼a 2142584
▼b (N$T) |
035 | |
▼a (OCoLC)1102478382 |
037 | |
▼a DE8DD0F1-F2A7-45ED-93BF-FAD5E156EC76
▼b OverDrive, Inc.
▼n http://www.overdrive.com |
040 | |
▼a EBLCP
▼b eng
▼c EBLCP
▼d TEFOD
▼d N$T
▼d OCLCO
▼d 248032 |
049 | |
▼a MAIN |
050 | 4 |
▼a QA76.73.P98 |
072 | 7 |
▼a COM
▼x 051360
▼2 bisacsh |
082 | 04 |
▼a 005.133
▼2 23 |
100 | 1 |
▼a Miller, Curtis. |
245 | 10 |
▼a Training Systems Using Python Statistical Modeling
▼h [electronic resource] :
▼b Explore Popular Techniques for Modeling Your Data in Python. |
260 | |
▼a Birmingham :
▼b Packt Publishing, Limited,
▼c 2019. |
300 | |
▼a 1 online resource (284 p.) |
500 | |
▼a Description based upon print version of record. |
500 | |
▼a The silhouette method |
505 | 0 |
▼a Cover; Title Page; Copyright and Credits; About Packt; Contributors; Table of Contents; Preface; Chapter 1: Classical Statistical Analysis; Technical requirements; Computing descriptive statistics; Preprocessing the data; Computing basic statistics; Classical inference for proportions; Computing confidence intervals for proportions; Hypothesis testing for proportions; Testing for common proportions; Classical inference for means; Computing confidence intervals for means; Hypothesis testing for means; Testing with two samples; One-way analysis of variance (ANOVA); Diving into Bayesian analysis |
505 | 8 |
▼a How Bayesian analysis worksUsing Bayesian analysis to solve a hit-and-run; Bayesian analysis for proportions; Conjugate priors for proportions; Credible intervals for proportions; Bayesian hypothesis testing for proportions; Comparing two proportions; Bayesian analysis for means; Credible intervals for means; Bayesian hypothesis testing for means; Testing with two samples; Finding correlations; Testing for correlation; Summary; Chapter 2: Introduction to Supervised Learning; Principles of machine learning; Checking the variables using the iris dataset; The goal of supervised learning |
505 | 8 |
▼a Training modelsIssues in training supervised learning models; Splitting data; Cross-validation; Evaluating models; Accuracy; Precision; Recall; F1 score; Classification report; Bayes factor; Summary; Chapter 3: Binary Prediction Models; K-nearest neighbors classifier; Training a kNN classifier; Hyperparameters in kNN classifiers; Decision trees; Fitting the decision tree; Visualizing the tree; Restricting tree depth; Random forests; Optimizing hyperparameters; Naive Bayes classifier; Preprocessing the data; Training the classifier; Support vector machines; Training a SVM; Logistic regression |
505 | 8 |
▼a Fitting a logit modelExtending beyond binary classifiers; Multiple outcomes for decision trees; Multiple outcomes for random forests; Multiple outcomes for Naive Bayes; One-versus-all and one-versus-one classification; Summary; Chapter 4: Regression Analysis and How to Use It; Linear models; Fitting a linear model with OLS; Performing cross-validation; Evaluating linear models; Using AIC to pick models; Bayesian linear models; Choosing a polynomial; Performing Bayesian regression; Ridge regression; Finding the right alpha value; LASSO regression; Spline interpolation |
505 | 8 |
▼a Using SciPy for interpolation2D interpolation; Summary; Chapter 5: Neural Networks; An introduction to perceptrons; Neural networks; The structure of a neural network; Types of neural networks; The MLP model; MLPs for classification; Optimization techniques; Training the network; Fitting an MLP to the iris dataset; Fitting an MLP to the digits dataset; MLP for regression; Summary; Chapter 6: Clustering Techniques; Introduction to clustering; Computing distances; Exploring the k-means algorithm; Clustering the iris dataset; Compressing images with k-means; Evaluating clusters; The elbow method |
520 | |
▼a This book will acquaint you with various aspects of statistical analysis in Python. You will work with different types of prediction models, such as decision trees, random forests and neural networks. By the end of this book, you will be confident in using various Python packages to train your own models for effective machine learning. |
590 | |
▼a Master record variable field(s) change: 050, 072, 082, 650 |
650 | 7 |
▼a COMPUTERS
▼x Programming Languages
▼x Python.
▼2 bisacsh |
650 | 0 |
▼a Python (Computer program language) |
650 | 0 |
▼a Graphical modeling (Statistics) |
655 | 4 |
▼a Electronic books. |
776 | 08 |
▼i Print version:
▼a Miller, Curtis
▼t Training Systems Using Python Statistical Modeling : Explore Popular Techniques for Modeling Your Data in Python
▼d Birmingham : Packt Publishing, Limited,c2019
▼z 9781838823733 |
856 | 40 |
▼3 EBSCOhost
▼u http://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=2142584 |
938 | |
▼a EBL - Ebook Library
▼b EBLB
▼n EBL5778835 |
938 | |
▼a EBSCOhost
▼b EBSC
▼n 2142584 |
990 | |
▼a 관리자 |
994 | |
▼a 92
▼b N$T |