Schweitzer Fachinformationen
Wenn es um professionelles Wissen geht, ist Schweitzer Fachinformationen wegweisend. Kunden aus Recht und Beratung sowie Unternehmen, öffentliche Verwaltungen und Bibliotheken erhalten komplette Lösungen zum Beschaffen, Verwalten und Nutzen von digitalen und gedruckten Medien.
A comprehensive presentation of the implementation of artificial intelligence and quantum computing technology in large-scale communication networks
Increasingly dense and flexible wireless networks require the use of artificial intelligence (AI) for planning network deployment, optimization, and dynamic control. Machine learning algorithms are now often used to predict traffic and network state in order to reserve resources for smooth communication with high reliability and low latency.
In Artificial Intelligence and Quantum Computing for Advanced Wireless Networks, the authors deliver a practical and timely review of AI-based learning algorithms, with several case studies in both Python and R. The book discusses the game-theory-based learning algorithms used in decision making, along with various specific applications in wireless networks, like channel, network state, and traffic prediction. Additional chapters include Fundamentals of ML, Artificial Neural Networks (NN), Explainable and Graph NN, Learning Equilibria and Games, AI Algorithms in Networks, Fundamentals of Quantum Communications, Quantum Channel, Information Theory and Error Correction, Quantum Optimization Theory, and Quantum Internet, to name a few.
The authors offer readers an intuitive and accessible path from basic topics on machine learning through advanced concepts and techniques in quantum networks. Readers will benefit from:
Perfect for network engineers, researchers, and graduate and masters students in computer science and electrical engineering, Artificial Intelligence and Quantum Computing for Advanced Wireless Networks is also an indispensable resource for IT support staff, along with policymakers and regulators who work in technology.
Savo G. Glisic is Research Professor at Worcester Polytechnic Institute, Massachusetts, USA. His research interests include network optimization theory, network topology control and graph theory, cognitive networks, game theory, artificial intelligence, and quantum computing technology.
Beatriz Lorenzo is Assistant Professor in the Department of Electrical and Computer Engineering at the University of Massachusetts Amherst, USA. Her research interests include the areas of communication networks, wireless networks, and mobile computing.
Preface, xiii
Part I Artificial Intelligence, 1
1 Introduction, 3
1.1 Motivation, 3
1.2 Book Structure, 5
2 Machine Learning Algorithms, 17
2.1 Fundamentals, 17
2.2 ML Algorithm Analysis, 37
3 Artificial Neural Networks, 55
3.1 Multi-layer Feedforward Neural Networks, 55
3.2 FIR Architecture, 60
3.3 Time Series Prediction, 68
3.4 Recurrent Neural Networks, 69
3.5 Cellular Neural Networks (CeNN), 81
3.6 Convolutional Neural Network (CoNN), 84
4 Explainable Neural Networks, 97
4.1 Explainability Methods, 99
4.2 Relevance Propagation in ANN, 103
4.3 Rule Extraction from LSTM Networks, 110
4.4 Accuracy and Interpretability, 112
5 Graph Neural Networks, 135
5.1 Concept of Graph Neural Network (GNN), 135
5.2 Categorization and Modeling of GNN, 144
5.3 Complexity of NN, 156
6 Learning Equilibria and Games, 179
6.1 Learning in Games, 179
6.2 Online Learning of Nash Equilibria in Congestion Games, 196
6.3 Minority Games, 202
6.4 Nash Q-Learning, 204
6.5 Routing Games, 211
6.6 Routing with Edge Priorities, 220
7 AI Algorithms in Networks, 227
7.1 Review of AI-Based Algorithms in Networks, 227
7.2 ML for Caching in Small Cell Networks, 237
7.3 Q-Learning-Based Joint Channel and Power Level Selection in Heterogeneous Cellular Networks, 243
7.4 ML for Self-Organizing Cellular Networks, 252
7.5 RL-Based Caching, 267
7.6 Big Data Analytics in Wireless Networks, 274
7.7 Graph Neural Networks, 279
7.8 DRL for Multioperator Network Slicing, 291
7.9 Deep Q-Learning for Latency-Limited Network Virtualization, 302
7.10 Multi-Armed Bandit Estimator (MBE), 317
7.11 Network Representation Learning, 327
Part II Quantum Computing, 361
8 Fundamentals of Quantum Communications, 363
8.1 Introduction, 363
8.2 Quantum Gates and Quantum Computing, 372
8.3 Quantum Fourier Transform (QFT), 386
9 Quantum Channel Information Theory, 397
9.1 Communication Over a Channel, 398
9.2 Quantum Information Theory, 401
9.3 Channel Description, 407
9.4 Channel Classical Capacities, 414
9.5 Channel Quantum Capacity, 431
9.6 Quantum Channel Examples, 437
10 Quantum Error Correction, 451
10.1 Stabilizer Codes, 458
10.2 Surface Code, 465
10.3 Fault-Tolerant Gates, 471
10.4 Theoretical Framework, 474
11 Quantum Search Algorithms, 499
11.1 Quantum Search Algorithms, 499
11.2 Physics of Quantum Algorithms, 510
12 Quantum Machine Learning, 543
12.1 QML Algorithms, 543
12.2 QNN Preliminaries, 547
12.3 Quantum Classifiers with ML: Near-Term Solutions, 550
12.4 Gradients of Parameterized Quantum Gates, 560
12.5 Classification with QNNs, 568
12.6 Quantum Decision Tree Classifier, 575
13 QC Optimization, 593
13.1 Hybrid Quantum-Classical Optimization Algorithms, 593
13.2 Convex Optimization in Quantum Information Theory, 601
13.3 Quantum Algorithms for Combinatorial Optimization Problems, 609
13.4 QC for Linear Systems of Equations, 614
13.5 Quantum Circuit, 625
13.6 Quantum Algorithm for Systems of Nonlinear Differential Equations, 628
14 Quantum Decision Theory, 637
14.1 Potential Enablers for Qc, 637
14.2 Quantum Game Theory (QGT), 641
14.3 Quantum Decision Theory (QDT), 665
14.4 Predictions in QDT, 676
15 Quantum Computing in Wireless Networks, 693
15.1 Quantum Satellite Networks, 693
15.2 QC Routing for Social Overlay Networks, 706
15.3 QKD Networks, 713
16 Quantum Network on Graph, 733
16.1 Optimal Routing in Quantum Networks, 733
16.2 Quantum Network on Symmetric Graph, 744
16.3 QWs, 747
16.4 Multidimensional QWs, 753
17 Quantum Internet, 773
17.1 System Model, 775
17.2 Quantum Network Protocol Stack, 789
References, 814
Index, 821
Owing to the increase in the density and number of different functionalities in wireless networks, there is an increasing need for the use of artificial intelligence (AI) in planning the network deployment, running their optimization, and dynamically controlling their operation. Machine learning (ML) algorithms are used to predict traffic and network state in order to reserve resources for smooth communication with high reliability and low latency in a timely fashion. Big data mining is used to predict customer behavior and pre-distribute (caching) the information content across the network in a timely fashion so that it can be efficiently delivered as soon as it is requested. Intelligent agents can search the Internet on behalf of the customer in order to find the best options when it comes to buying any product online. This book reviews ML-based algorithms with a number of case studies supported by Python and R programs. It discusses the learning algorithms used in decision making based on game theory and a number of specific applications in wireless networks such as channel, network state, and traffic prediction.
We begin the book with a comprehensive survey of AI learning algorithms. These algorithms are used in the prediction of the network parameters for efficient network slicing, customer behavior for content caching across the network, or for efficient network control and management. Subsequently, we focus on network applications with an emphasis on AI-based learning algorithms used for reaching equilibria in games used among different parties in a variety of new business models in communication networks. This includes competition between network operators, service providers, or even users in dynamic network architectures of user-provided networks.
The book also covers in detail a number of specific applications of AI for dynamic readjusting network behavior based on the observation of its state, traffic variation, and user behavior. This includes channel and power level selection in cellular networks, network self-organization, proactive caching, big data learning, graph neural network (GNN), and multi-armed bandit estimators.
Why quantum computing? The ever-reducing transistor size following Moore's law is approaching the point where quantum effects predominate in transistor operation. This specific trend implies that quantum effects become unavoidable, hence making research on quantum computing (QC) systems an urgent necessity. In fact, a quantum annealing chipset is already commercially available from D-Wave1.
Apart from the quantum annealing architecture, gate-based architecture, which relies on building computational blocks using quantum gates in a similar fashion to classical logic gates, is attracting increasing attention due to the recent advances in quantum stabilizer codes, which are capable of mitigating the de-coherence effects encountered by quantum circuits. In terms of implementation, IBM has initially produced 53-qubits quantum computer [1] and plans to have 1-million qubits by 2030 [2]. D-Wave Two 512 qubit processors [3] are built in Google and NASA quantum computer. With this recent developments, Quantum computing has become a commercial reality and it may be used in wireless communications systems in order to speed up specific processes due to its inherent parallelization capabilities.
Whereas a classical bit may adopt the values 0 or 1, a quantum bit, or qubit, may have the values |0>, |1>, or any superposition of the two, where the notation |> is the column vector of a quantum state. If two qubits are used, then the composite quantum state may have the values |00>, |01>, |10>, and |11> simultaneously. In general, by employing b bits in a classical register, one out of b2 combinations is represented at any time. By contrast, in a quantum register associated with b qubits, the composite quantum state may be found in a superposition of all b2 values simultaneously. Therefore, applying a quantum operation to the quantum register would result in altering all b2 values at the same time. This represents the parallel processing capability of quantum computing.
In addition to superior computing capabilities, multiple quantum algorithms have been proposed, which are capable of outperforming their classical counterparts in the same categories of problems, by either requiring fewer computational steps, or by finding a better solution to the specific problem. In this book, we will focus on the employment of quantum algorithms in classical communication systems, which is nowadays referred to as quantum-assisted communications.
In the following sections, we revisit the ML methods in the context of quantum-assisted algorithms for ML and the quantum machine learning (QML) framework. Quantum principles based on emerging computing technologies will bring in entirely new modes of information processing. An overview of supervised, unsupervised, and reinforcement learning (RL) methods for QML is presented in this segment of the book.
Currently, 5G networks have entered into the commercialization phase, which makes it appropriate to launch a strong effort to conceptualize the future vision of the next generation of wireless networks. The increasing size, complexity, services, and performance demands of communication networks necessitate planning and consultation for envisioning new technologies to enable and harmonize future heterogeneous networks. An overwhelming interest in AI methods is seen in recent years, which has motivated the provision of essential intelligence to 5G networks. However, this provision is limited to the performance of different isolated tasks of optimization, control, and management. The recent success of quantum-assisted and data-driven learning methods in communication networks has led to their candidature as enablers of future heterogeneous networks. This section reviews a novel framework for 6G/7G networks, where quantum-assisted ML and QML are proposed as the core enablers along with some promising communication technology innovations.
The relevance of the research fields integrated throughout this book can be easily recognized within the National Science Foundation (NSF) list of research priorities in science and technology: These 10 areas specified by NSF include (i) AI and ML; (ii) high performance computing, semiconductors, and advanced computer hardware; (iii) quantum computing and information systems; (iv) robotics, automation, and advanced manufacturing; (v) natural or anthropogenic disaster prevention; (vi) advanced communications technology; (vii) biotechnology, genomics, and synthetic biology; (viii) cybersecurity, data storage, and data management technologies; (ix) advanced energy; and (x) materials science, engineering, and exploration relevant to other key technology areas. The 10 areas would be revisited every four years.
The first part of the book covers selected topics in ML, and the second part presents a number of topics from QC relevant for networking.
Chapter 2 (Machine Learning Algorithms): This chapter presents an introductory discussion of many basic ML algorithms that are often used in practice and not necessary directly related to networking problems. However, they will present a logical basis for developing more sophisticated algorithms that are used nowadays to efficiently solve various problems in this field. These algorithms include linear regression, logistic regression, decision tree (regression trees vs. classification trees), and working with decision trees [4] in R and Python. In this chapter, we answer the questions: What is bagging? What is random forest? What is boosting? Which is more powerful: GBM or XGBoost? We also explain the basics of working in R and Python with GBM, XGBoost, SVM (support vector machine), Naive Bayes, kNN, K-means, random forest, dimensionality reduction algorithms [5, 6], gradient boosting algorithms, GBM, XGBoost, LightGBM, and CatBoost [7, 8].
Chapter 3 (Artificial Neural Networks): We are witnessing the rapid, widespread adoption of AI [9] in our daily life, which is accelerating the shift toward a more algorithmic society. Our focus is on reviewing the unprecedented new opportunities opened up by using AI in deploying and optimization of communication networks. In this chapter, we will discuss the basis of artificial neural networks (ANNs) [10] including multilayer neural networks, training and backpropagation, finite-impulse response (FIR) architecture spatial temporal representations, derivation of temporal backpropagation, applications in time series prediction, auto-regressive linear prediction, nonlinear prediction, adaptation and iterated predictions as well as multiresolution FIR neural-network-based learning algorithm applied to network traffic prediction. Traffic prediction is important for timely reconfiguration of the network topology or traffic rerouting to avoid congestion or network slicing.
Chapter 4 (Explainable NN): Even with the advancements of AI described in the previous chapter, a key impediment to the use of AI-based systems is that they often lack transparency. Indeed, the black-box nature of these systems allows powerful predictions, but they cannot be directly explained. This problem has...
Dateiformat: ePUBKopierschutz: Adobe-DRM (Digital Rights Management)
Systemvoraussetzungen:
Das Dateiformat ePUB ist sehr gut für Romane und Sachbücher geeignet – also für „fließenden” Text ohne komplexes Layout. Bei E-Readern oder Smartphones passt sich der Zeilen- und Seitenumbruch automatisch den kleinen Displays an. Mit Adobe-DRM wird hier ein „harter” Kopierschutz verwendet. Wenn die notwendigen Voraussetzungen nicht vorliegen, können Sie das E-Book leider nicht öffnen. Daher müssen Sie bereits vor dem Download Ihre Lese-Hardware vorbereiten.Bitte beachten Sie: Wir empfehlen Ihnen unbedingt nach Installation der Lese-Software diese mit Ihrer persönlichen Adobe-ID zu autorisieren!
Weitere Informationen finden Sie in unserer E-Book Hilfe.