Networked control systems are increasingly ubiquitous today, with applications ranging from vehicle communication and adaptive power grids to space exploration and economics. The optimal design of such systems presents major challenges, requiring tools from various disciplines within applied mathematics such as decentralized control, stochastic control, information theory, and quantization. A thorough, self-contained book, Stochastic Networked Control Systems: Stabilization and Optimization under Information Constraints aims to connect these diverse disciplines with precision and rigor, while conveying design guidelines to controller architects. Unique in the literature, it lays a comprehensive theoretical foundation for the study of networked control systems, and introduces an array of concrete tools for work in the field. Salient features included: · Characterization, comparison and optimal design of information structures in static and dynamic teams. Operational, structural and topological properties of information structures in optimal decision making, with a systematic program for generating optimal encoding and control policies. The notion of signaling, and its utilization in stabilization and optimization of decentralized control systems. · Presentation of mathematical methods for stochastic stability of networked control systems using random-time, state-dependent drift conditions and martingale methods. · Characterization and study of information channels leading to various forms of stochastic stability such as stationarity, ergodicity, and quadratic stability; and connections with information and quantization theories. Analysis of various classes of centralized and decentralized control systems. · Jointly optimal design of encoding and control policies over various information channels and under general optimization criteria, including a detailed coverage of linear-quadratic-Gaussian models. · Decentralized agreement and dynamic optimization under information constraints. This monograph is geared toward a broad audience of academic and industrial researchers interested in control theory, information theory, optimization, economics, and applied mathematics. It could likewise serve as a supplemental graduate text. The reader is expected to have some familiarity with linear systems, stochastic processes, and Markov chains, but the necessary background can also be acquired in part through the four appendices included at the end. · Characterization, comparison and optimal design of information structures in static and dynamic teams. Operational, structural and topological properties of informationstructures in optimal decision making, with a systematic program for generating optimal encoding and control policies. The notion of signaling, and its utilization in stabilization and optimization of decentralized control systems. · Presentation of mathematical methods for stochastic stability of networked control systems using random-time, state-dependent drift conditions and martingale methods. · Characterization and study of information channels leading to various forms of stochastic stability such as stationarity, ergodicity, and quadratic stability; and connections with information and quantization theories. Analysis of various classes of centralized and decentralized control systems. · Jointly optimal design of encoding and control policies over various information channels and under general optimization criteria, including a detailed coverage of linear-quadratic-Gaussian models. · Decentralized agreement and dynamic optimization under information constraints. This monograph is geared toward a broad audience of academic and industrial researchers interest