Introduction to Machine Learning Physics

橋本 幸士(編)

橋本 幸士(編)

定価 4,400 円(本体 4,000 円+税)

A5判/192ページ
刊行日:2025年10月01日
ISBN:978-4-254-13154-3 C3042

ネット書店で購入する amazon e-hon 紀伊國屋書店 丸善ジュンク堂書店 Honya Club Rakutenブックス くまざわ書店

書店の店頭在庫を確認する 紀伊國屋書店

内容紹介

『学習物理学入門』の英語版。An introductory textbook that examines the interplay between physics and AI/machine learning. Aimed at physics students, it provides a smooth entry into machine learning and explores the collaborative relationship between the two fields. [language: English]

編集部から

[Machine Learning Physics AI Bot]
This book implements a system where you can learn alongside large language models (GPTs). The AI, primarily developed by author Akiyoshi Sannai, engages in dialogue with readers based on the book's content. We encourage you to learn about AI together with AI. Please access the AI using the URL below.

https://chatgpt.com/g/g-Ch3xAdHpq-xue-xi-wu-li-jie-shuo-bot

• Please note that we may not be able to respond to inquiries about this AI bot.
• This AI bot may terminate its service without prior notice.
• Important: The AI bot's initial interface is in Japanese, but it can respond to any question in English. The AI bot was built on the knowledge of the Japanese edition of this book, so it may refer to page numbers or other details that do not exactly match this English edition.

[Review]
'From Newton to neural nets, via entropy, symmetry, and path integrals: this book provides a wonderful tour through the surprising parallels that link physics and machine learning.'
― Prof. David Tong, University of Cambridge

'Physics and machine learning share deep conceptual connections. This textbook will guide the reader through profound insights and exciting advances at the emerging intersection of these two fields.'
― Prof. Jesse Thaler, MIT/Director of IAIFI

目次

[Table of Contents]

Preface
Introduction

A Machine Learning and Physics

A1. Linear Models
 A1.1 Least Squares Method and Linear Regression
  A1.1.1 Least Squares Method
  A1.1.2 Convex Functions
  A1.1.3 Conditions for Convexity of Multivariate Functions
  A1.1.4 Linear Models
  A1.1.5 Continuation of Least Squares Method
 A1.2 Entropy
  A1.2.1 Probability
  A1.2.2 Shannon Entropy
  A1.2.3 Relative Entropy and KL Divergence
  A1.2.4 Jensen's Inequality
  A1.2.5 Gaussian Distribution
 A1.3 Maximum Likelihood Estimation
  A1.3.1 Likelihood Function
  A1.3.2 Maximum Likelihood from KL Divergence
 A1.4 Generalized Linear Models
  A1.4.1 Binary Classification and Logistic Regression
  A1.4.2 Origin of Cross-Entropy
 A1.5 Classification of Machine Learning
 A1.6 Generalization, Overfitting and Underfitting
 A1.7 Random Numbers
  A1.7.1 What are Random Numbers
  A1.7.2 Uniform Random Numbers
  A1.7.3 Gaussian Random Numbers

A2. Neural Networks (NN)
 A2.1 Neural Networks
 A2.2 Data Representation
  A2.2.1 Vectorization of Images
  A2.2.2 One-Hot Representation
 A2.3 Fully Connected Neural Networks with General Number of Layers
 A2.4 Gradient Descent Method
 A2.5 Activation Functions and Their Derivatives
 A2.6 Backpropagation
 A2.7 Gradient Vanishing Problem

A3. Symmetry and Machine Learning: Convolution and Equivariant NN
 A3.1 Equivariance and Convolutional Neural Networks
 A3.2 Image Filters
 A3.3 Convolutional Layer
  A3.3.1 Two-Dimensional Convolution
  A3.3.2 Pooling
 A3.4 Group Theory and Symmetry
 A3.5 Symmetry and Equivariance
  A3.5.1 Ways to Incorporate Symmetry
  A3.5.2 Group Equivariant Neural Networks
  A3.5.3 Inductive Bias
  A3.5.4 Gauge Symmetry and Neural Networks
  
A4. Classical Mechanics and Machine Learning: Neural Networks and Differential Equations
 A4.1 Fundamental Equations of Physics and Machine Learning
  A4.1.1 The Role of Differential Equations
  A4.1.2 Embedding Physics Problems into Machine Learning
 A4.2 Physics-Informed Neural Networks (PINN)
 A4.3 Viewing Neural Networks as Differential Equations
  A4.3.1 Methods for Handling Differential Equations in Machine Learning
  A4.3.2 Locality of NN
  A4.3.3 ResNet and Differential Equations
  A4.3.4 Locality within Layers and Convolutional NN
 A4.4 Representation of Specific Equations of Motion by NN
  A4.4.1 Example of a Particle in a Potential
  A4.4.2 Hamiltonian Systems

A5. Quantum Mechanics and Machine Learning: Neural Network Wave Functions
 A5.1 Quantum Mechanics and Eigenvalue Problems
 A5.2 Quantum Many-Body Problems on Lattices
 A5.3 Variational Method and Trial Functions
 A5.4 Neural Network Wave Functions in Small Quantum Systems
  A5.4.1 Analytical Solution of the Two-Site Transverse-Field Ising Model
  A5.4.2 Approximate Solution Using Neural NetworkWave Functions
 A5.5 Neural Network Wave Functions in Larger Quantum Systems
  A5.5.1 Exact Numerical Solution Using the Exact Diagonalization Method
  A5.5.2 Approximate Numerical Solution Using Neural Network Wave Function
 A5.6 Future Prospects
 
B Machine Learning Models and Physics

B1. Transformer
 B1.1 Words and Embedding Vectors
  B1.1.1 Use Theory of Meaning and Embedding
  B1.1.2 Search from Key-Value Store and Attention Mechanism
  B1.1.3 Transformer Architecture
 B1.2 Transformers in NLP and Computer Vision
  B1.2.1 GPT
  B1.2.2 Vision Transformer
  
B2. Diffusion Models and Path Integrals
 B2.1 Principles of Diffusion Models
  B2.1.1 The Idea of Diffusion Models
  B2.1.2 Diffusion Models and Langevin Equation
  B2.1.3 Sampling Process of Diffusion Models
  B2.1.4 Training of Diffusion Models
  B2.1.5 Probability Flow ODE
 B2.2 Path Integral Quantization
 B2.3 Path Integral Formulation of Diffusion Models
  B2.3.1 Derivation of the Reverse Process
  B2.3.2 Derivation of Loss Function for Diffusion Model Training
  B2.3.3 Probability Flow and Classical Limit

B3. Mechanism Behind Machine Learning: Statistical Mechanical Approach
 B3.1 Infinite-Width DNN: Signal Propagation
  B3.1.1 Spin Model
  B3.1.2 Macroscopic Laws of Signal Propagation
  B3.1.3 Mean Field Theory and Order-to-Chaos Phase Transition
  B3.1.4 Macroscopic Law of Backpropagation
  B3.1.5 Vanishing and Exploding Gradient Problem as Phase Transition
  B3.1.6 Connection with Kernel Methods
 B3.2 Infinite-Width DNN Model: Learning Regimes
  B3.2.1 NTK Regime
  B3.2.2 μP
 B3.3 Linear Regression Model
  B3.3.1 Generalization Error in Over-Parameterized Models
  B3.3.2 Typical Evaluation of Generalization Error
  
B4. Large Language Models and Science
 B4.1 Large Language Models
  B4.1.1 Next Word Prediction
  B4.1.2 Training of Large Language Models
 B4.2 Applications of Large Language Models
  B4.2.1 Arithmetic Capabilities of Large Language Models
  B4.2.2 Proof Capabilities of Large Language Models
  B4.2.3 Cubism in Mathematics

Afterword
Index

執筆者紹介

[Contributors]
Koji Hashimoto is at the Graduate School of Science, Kyoto University, Japan
Akio Tomiya is at the School of Arts and Sciences, Tokyo Woman's Christian University, Japan
Ryui Kaneko is at the Faculty of Science and Technology, Sophia University, Japan
Masato Taki is at the Graduate School of Artificial Intelligence and Science, Rikkyo University, Japan
Yuji Hirono is at the Institute of Systems and Information Engineering, University of Tsukuba, Japan
Ryo Karakida is at the Artificial Intelligence Research Center, National Institute of Advanced Industrial Science and Technology, Japan
Akiyoshi Sannai is at the Graduate School of Science, Kyoto University, Japan

関連情報

ジャンル一覧

ジャンル一覧

  • Facebook
  • X
  • 「愛読者の声」 ご投稿はこちら 「愛読者の声」 ご投稿はこちら
  • EBSCO eBooks
  • eBook Library