幫助中心 | 我的帳號 | 關於我們

機器學習統計學(影印版)(英文版)

  • 作者:(印)普拉塔普·丹格迪
  • 出版社:東南大學
  • ISBN:9787564177553
  • 出版日期:2018/08/01
  • 裝幀:平裝
  • 頁數:426
人民幣:RMB 98 元      售價:
放入購物車
加入收藏夾

內容大鋼
    機器學習所涉及的複雜統計學知識困擾了很多開發者。知曉統計學知識可以幫助你為給定的問題構建強壯的機器學習優化模型。
    普拉塔普·丹格迪著的《機器學習統計學(影印版)(英文版)》將教你機器學習所需的實現複雜統計計算的相關內容,可以從中獲得監督學習、無監督學習、強化學習等背後的統計學知識。你將看到討論機器學習相關統計學內容的真實案例並熟悉它們。還能學到用於實現建模、調參、回歸、分類、密度採集、向量處理、矩陣等的相關程序。
    學完本書,你會掌握機器學習所需的統計學知識。並且能夠將所學新技能應用於任何行業問題。

作者介紹
(印)普拉塔普·丹格迪

目錄
Preface
Chapter 1: Journey from Statistics to Machine Learning
  Statistical terminology for model building and validation
  Machine learning
  Major differences between statistical modeling and machine learning
  Steps in machine learning model development and deployment
  Statistical fundamentals and terminology for model building andvalidation
  Bias versus variance trade-off
  Train and test data
  Machine learning terminology for model building and validation
  Linear regression versus gradient descent
  Machine learning losses
  When to stop tuning machine learning models
  Train, validation, and test data
  Cross-validation
  Grid search
  Machine learning model overview
  Summary
Chapter 2: Parallelism of Statistics and Machine Learning
  Comparison between regression and machine learning models
  Compensating factors in machine learning models
  Assumptions of linear regression
  Steps applied in linear regression modeling
  Example of simple linear regression from first principles
  Example of simple linear regression using the wine quality data
  Example of multilinear regression - step-by-step methodology of model
  building
  Backward and forward selection
  Machine learning models - ridge and lasso regression
  Example of ridge regression machine learning
  Example of lasso regression machine learning model
  Regularization parameters in linear regression and ridge/lasso regression
  Summary
Chapter 3: Logistic Regression Versus Random Forest
  Maximum likelihood estimation
  Logistic regression - introduction and advantages
  Terminology involved in logistic regression
  Applying steps in logistic regression modeling
  Example of logistic regression using German credit data
  Random forest
  Example of random forest using German credit data
  Grid search on random forest
  Variable importance plot
  Comparison of logistic regression with random forest
  Summary
Chapter 4: Tree-Based Machine Learning Models
  Introducing decision tree classifiers
  Terminology used in decision trees
  Decision tree working methodology from first principles
  Comparison between logistic regression and decision trees

  Comparison of error components across various styles of models
  Remedial actions to push the model towards the ideal region
  HR attrition data example
  Decision tree classifier
  Tuning class weights in decision tree classifier
  Bagging classifier
  Random forest classifier
  Random forest classifier - grid search
  AdaBoost classifier
  Gradient boosting classifier
  Comparison between AdaBoosting versus gradient boosting
  Extreme gradient boosting - XGBoost classifier
  Ensemble of ensembles - model stacking
  Ensemble of ensembles with different types of classifiers
  Ensemble of ensembles with bootstrap samples using a single type of
  classifier
  Summary
Chapter 5: K-Nearest Neighbors and Naive Bayes
  K-nearest neighbors
  KNN voter example
  Curse of dimensionality
  Curse of dimensionality with 1D, 2D, and 3D example
  KNN classifier with breast cancer Wisconsin data example
  Tuning of k-value in KNN classifier
  Naive Bayes
  Probability fundamentals
  Joint probability
  Understanding Bayes theorem with conditional probability
  Naive Bayes classification
  Laplace estimator
  Naive Bayes SMS spam classification example
  Summary
Chapter 6: Support Vector Machines and Neural Networks
  Support vector machines working principles
  Maximum margin classifier
  Support vector classifier
  Support vector machines
  Kernel functions
  SVM multilabel classifier with letter recognition data example
  Maximum margin classifier - linear kernel
  Polynomial kernel
  RBF kernel
  Artificial neural networks -ANN
  Activation functions
  Forward propagation and backpropagation
  Optimization of neural networks
  Stochastic gradient descent - SGD
  Momentum
  Nesterov accelerated gradient - NAG
  Adagrad

  Adadelta
  RMSprop
  Adaptive moment estimation - Adam
  Limited-memory broyden-fletcher-goldfarb-shanno - L-BFGS
  optimization algorithm
  Dropout in neural networks
  ANN classifier applied on handwritten digits using scikit-learn
  Introduction to deep learning
  Solving methodology
  Deep learning software
  Deep neural network classifier applied on handwritten digits using Keras
  Summary
Chapter 7: Recommendation Engines
  Content-based filtering
  Cosine similarity
  Collaborative filtering
  Advantages of collaborative filtering over content-based filtering
  Matrix factorization using the alternating least squares algorithm for
  collaborative filtering
  Evaluation of recommendation engine model
  Hyperparameter selection in recommendation engines using grid search
  Recommendation engine application on movie lens data
  User-user similarity matrix
  Movie-movie similarity matrix
  Collaborative filtering using ALS
  Grid search on collaborative filtering
  Summary
Chapter 8: Unsupervised Learning
  K-means clustering
  K-means working methodology from first principles
  Optimal number of clusters and cluster evaluation
  The elbow method
  K-means clustering with the iris data example
  Principal component analysis - PCA
  PCA working methodology from first principles
  PCA applied on handwritten digits using scikit-learn
  Singular value decomposition - SVD
  SVD applied on handwritten digits using scikit-learn
  Deep auto encoders
  Model building technique using encoder-decoder architecture
  Deep auto encoders applied on handwritten digits using Keras
  Summary
Chapter 9: Reinforcement Learning
  Introduction to reinforcement learning
  Comparing supervised, unsupervised, and reinforcement learning in detail
  Characteristics of reinforcement learning
  Reinforcement learning basics
  Category 1 - value based
  Category 2 - policy based
  Category 3 - actor-critic

  Category 4 - model-free
  Category 5 - model-based
  Fundamental categories in sequential decision making
  Markov decision processes and Bellman equations
  Dynamic programming
  Algorithms to compute optimal policy using dynamic programming
  Grid world example using value and policy iteration algorithms with basic Python
  Monte Carlo methods
  Comparison between dynamic programming and Monte Carlo methods
  Key advantages of MC over DP methods
  Monte Carlo prediction
  The suitability of Monte Carlo prediction on grid-world problems
  Modeling Blackjack example of Monte Carlo methods using Python
  Temporal difference learning
  Comparison between Monte Carlo methods and temporal difference
  learning
  TD prediction
  Driving office example for TD learning
  SARSA on-policy TD control
  Q-learning - off-policy TD control
  Cliff walking example of on-policy and off-policy of TD control
  Applications of reinforcement learning with integration of machine
  learning and deep learning
  Automotive vehicle control - self-driving cars
  Google DeepMind's AlphaGo
  Robo soccer
  Further reading
  Summary
Index

  • 商品搜索:
  • | 高級搜索
首頁新手上路客服中心關於我們聯絡我們Top↑
Copyrightc 1999~2008 美商天龍國際圖書股份有限公司 臺灣分公司. All rights reserved.
營業地址:臺北市中正區重慶南路一段103號1F 105號1F-2F
讀者服務部電話:02-2381-2033 02-2381-1863 時間:週一-週五 10:00-17:00
 服務信箱:bookuu@69book.com 客戶、意見信箱:cs@69book.com
ICP證:浙B2-20060032