'This book is an exciting addition to the literature on machine learning and graphical models. What makes it unique and interesting is that it provides a unified treatment of machine learning and related fields through graphical models, a framework of growing importance and popularity. Another feature of this book lies in its smooth transition from traditional artificial intelligence to modern machine learning. The book is well-written and truly pleasant to read. I believe that it will appeal to students and researchers with or without a solid mathematical background.' Zheng-Hua Tan, Aalborg University, Denmark
Preface; Part I. Inference in Probabilistic Models: 1. Probabilistic reasoning; 2. Basic graph concepts; 3. Belief networks; 4. Graphical models; 5. Efficient inference in trees; 6. The junction tree algorithm; 7. Making decisions; Part II. Learning in Probabilistic Models: 8. Statistics for machine learning; 9. Learning as inference; 10. Naive Bayes; 11. Learning with hidden variables; 12. Bayesian model selection; Part III. Machine Learning: 13. Machine learning concepts; 14. Nearest neighbour classification; 15. Unsupervised linear dimension reduction; 16. Supervised linear dimension reduction; 17. Linear models; 18. Bayesian linear models; 19. Gaussian processes; 20. Mixture models; 21. Latent linear models; 22. Latent ability models; Part IV. Dynamical Models: 23. Discrete-state Markov models; 24. Continuous-state Markov models; 25. Switching linear dynamical systems; 26. Distributed computation; Part V. Approximate Inference: 27. Sampling; 28. Deterministic approximate inference; Appendix. Background mathematics; Bibliography; Index.
Barber, David
David Barber is Reader in Information Processing in the Department of Computer Science, University College London.