LEADER 11663nam 2200541 450 001 9910555012703321 005 20221206172137.0 010 $a1-119-70591-6 010 $a1-119-70587-8 010 $a1-119-70592-4 024 7 $a10.1002/9781119705925 035 $a(CKB)4100000010871042 035 $a(MiAaPQ)EBC6317471 035 $a(CaBNVSL)mat09292525 035 $a(IDAMS)0b0000648d5918e0 035 $a(IEEE)9292525 035 $a(EXLCZ)994100000010871042 100 $a20210105d2020 uy 101 0 $aeng 135 $aurcnu|||||||| 181 $ctxt$2rdacontent 182 $cc$2rdamedia 183 $acr$2rdacarrier 200 10$aCognitive modeling of human memory and learning $ea non-invasive brain-computer interfacing approach /$fLidia Ghosh, Artificial Intelligence Lab., Dept. of Electronics and Tele-Communication Engineering, Amit Konar, Artificial Intelligence Lab., Dept. of Electronics and Tele-Communication Engineering, Pratyusha Rakshit, Artificial Intelligence Lab., Dept. of Electronics and Tele-Communication Engineering 210 1$aHoboken, New Jersey :$cWiley,$d[2020] 210 2$a[Piscataqay, New Jersey] :$cIEEE Xplore,$d[2020] 215 $a1 online resource (275 pages) 311 $a1-119-70586-X 320 $aIncludes bibliographical references and index. 327 $aChapter 1: Introduction to Human Memory and Learning Models -- 1.1 Introduction 2 -- 1.2 Philosophical Contributions to Memory Research 4 -- 1.2.1 Atkinson and Shiffrin's Model 4 -- 1.2.2 Tveter's Model 6 -- 1.2.3 Tulving's model 6 -- 1.2.4 The Parallel and Distributed Processing (PDP) Approach 8 -- 1.2.5 Procedural and Declarative Memory 9 -- 1.3 Brain-theoretic Interpretation of Memory Formation 11 -- 1.3.1 Coding for Memory 11 -- 1.3.2 Memory Consolidation 13 -- 1.3.3 Location of stored Memories 16 -- 1.3.4 Isolation of Information in Memory 16 -- 1.4 Cognitive Maps 17 -- 1.5 Neural Plasticity 18 -- 1.6 Modularity 19 -- 1.7 The cellular Process behind STM Formation 20 -- 1.8 LTM Formation 21 -- 1.9 Brain Signal Analysis in the Context of Memory and Learning 22 -- 1.9.1 Association of EEG alpha and theta band with memory performances 22 -- 1.9.2 Oscillatory beta and gamma frequency band activation in STM performance 26 -- 1.9.3 Change in EEG band power with changing working memory load 26 -- 1.9.4 Effects of Electromagnetic field on the EEG response of Working Memory 29 -- 1.9.5 EEG Analysis to discriminate focused attention and WM performance 30 -- 1.9.6 EEG power changes in memory repetition effect 31 -- 1.9.7 Correlation between LTM Retrieval and EEG features 34 -- 1.9.8 Impact of math anxiety on WM response: An EEG study 37 -- 1.10 Memory Modelling by Computational Intelligence Techniques 38 -- 1.11 Scope of the Book 43 -- References 47 -- Chapter 2: Working Memory Modeling Using Inverse Fuzzy Relational Approach -- 2.1 Introduction 56 -- 2.2 Problem Formulation and Approach 59 -- 2.2.1 Independent Component Analysis as a Source Localization Tool 61 -- 2.2.2 Independent Component Analysis vs Principal Component Analysis 62 -- 2.2.3 Feature Extraction 63 -- 2.2.4 Phase 1: WM Modeling 64 -- 2.2.4.1 Step I: WM modeling of subject using EEG signals during full face encoding and recall from specific part of same face 65 -- 2.2.4.2 Step II: WM modeling of subject using EEG signals during full face encoding and recall from all parts of same face 68. 327 $a2.2.5 Phase 2: WM Analysis 69 -- 2.2.6 Finding Max-Min Compositional of Weight Matrix 70 -- 2.3 Experimental Results and Performance Analysis 75 -- 2.3.1 Experimental Set-up 75 -- 2.3.2 Source Localization using e-LORETA 78 -- 2.3.3 Pre-processing 79 -- 2.3.4 Selection of EEG Features 80 -- 2.3.5 WM Model Consistency across Partial Face Stimuli 81 -- 2.3.6 Inter-person Variability of Weight Matrix W 85 -- 2.3.7 Variation in Imaging Attributes 87 -- 2.3.8 Comparative Analysis with existing Fuzzy Inverse Relations 87 -- 2.4 Discussion 88 -- 2.5 Conclusion 89 -- References 90 -- Chapter 3: Short-Term Memory Modeling in Shape-Recognition Task by Type-2 Fuzzy Deep Brain Learning -- 3.1 Introduction 98 -- 3.2 System Overview 101 -- 3.3 Brain Functional Mapping using Type-2 Fuzzy DBLN 107 -- 3.3.1 Overview of Type-2 Fuzzy Sets 107 -- 3.3.2 Type-2 Fuzzy Mapping and Parameter Adaptation by Perceptron-like Learning 108 -- 3.3.2.1 Construction of the Proposed Interval Type-2 Fuzzy Membership Function 109 -- 3.3.2.2 Construction of IT2FS Induced Mapping Function 110 -- 3.3.2.3 Secondary Membership Function Computation of Proposed GT2FS 112 -- 3.3.2.4 Proposed General Type-2 Fuzzy Mapping 114 -- 3.3.3 Perceptron-like Learning for Weight Adaptation 115 -- 3.3.4 Training of the Proposed Shape-Reconstruction Algorithm 116 -- 3.3.5 The Test Phase of the Memory Model 118 -- 3.4 Experiments and Results 118 -- 3.4.1 Experimental Set-up 118 -- 3.4.2 Experiment 1: Validation of the STM Model with respect to Error Metric 121 -- 3.4.3 Experiment 2: Similar Encoding by a Subject for Similar Input Object-Shapes 122 -- 3.4.4 Experiment 3: Study of Subjects' Learning Ability with Increasing Complexity in Object Shape 123 -- 3.4.5 Experiment 4: Convergence Time of the Weight Matrix G for Increased Complexity of the Input Shape Stimuli 124 -- 3.4.6 Experiment 5: Abnormality in G matrix for the subjects with Brain Impairment 125 -- 3.5 Biological Implications 126 -- 3.6 Performance Analysis 128. 327 $a3.6.1 Performance Analysis of the Proposed T2FS Methods 128 -- 3.6.2 Computational Performance Analysis of the Proposed T2FS Methods 130 -- 3.6.3 Statistical Validation using Wilcoxon Signed-Rank Test 130 -- 3.6.4 Optimal Parameter Selection and Robustness Study 131 -- 3.7 Conclusions 133 -- References 135 -- Chapter 4: EEG Analysis for Subjective Assessment of Motor Learning Skill in Driving Using Type-2 Fuzzy Reasoning -- 4.1 Introduction 142 -- 4.2 System Overview 144 -- 4.2.1 Rule Design to determine the degree of learning 145 -- 4.2.2 Single Trial Detection of Brain Signals 148 -- 4.2.2.1 Feature Extraction 149 -- 4.2.2.2 Feature Selection 149 -- 4.2.2.3 Classification 150 -- 4.2.3 Type-2 Fuzzy Reasoning 151 -- 4.2.4 Training and Testing of the Classifiers 151 -- 4.3 Determining Type and Degree of Learning by Type-2 Fuzzy Reasoning 151 -- 4.3.1 Preliminaries on IT2FS and GT2FS 153 -- 4.3.2 Proposed Reasoning Method 1: CIT2FS based Reasoning 153 -- 4.3.3 Computation of Percentage Normalized Degree of Learning 155 -- 4.3.4 Optimal ? Selection in IT2FS Reasoning 156 -- 4.3.5 Proposed Reasoning Method 2: Triangular Vertical Slice Based CGT2FS Reasoning 156 -- 4.3.6 Proposed Reasoning Method 3: CGT2FS Reasoning with Gaussian Secondary Membership Function (MF) 158 -- 4.4 Experiments and Results 162 -- 4.4.1 The Experimental set-up 162 -- 4.4.2 Stimulus Presentation 163 -- 4.4.3 Experiment 1: Source Localization using eLORETA 163 -- 4.4.4 Experiment 2: Validation of the Rules 164 -- 4.4.5 Experiment 3: Pre-processing and Artifact Removal using ICA 165 -- 4.4.6 Experiment 4: N400 Old/New Effect Observation over the Successive Trials 167 -- 4.4.7 Experiment 5: Selection of the Discriminating EEG Features using PCA 168 -- 4.5 Performance Analysis and Statistical Validation 169 -- 4.5.1 Performance Analysis of the LSVM Classifiers 169 -- 4.5.2 Robustness Study 170 -- 4.5.3 Performance Analysis of the Proposed T2FS Reasoning Methods 170 -- 4.5.4 Computational Performance Analysis of the Proposed T2FS Reasoning Methods 171. 327 $a4.5.5 Statistical Validation using Wilcoxon Signed-Rank Test 172 -- 4.6 Conclusion 173 -- References 173 -- Chapter 5: EEG Analysis to Decode Human Memory Responses in Face Recognition Task Using Deep LSTM Network -- 5.1 Introduction 182 -- 5.2 CSP Modeling 186 -- 5.2.1 The Standard CSP Algorithm 186 -- 5.2.2 The Proposed CSP Algorithm 187 -- 5.3 Proposed LSTM Classifier with Attention Mechanism 189 -- 5.4 Experiment and Results 195 -- 5.4.1 The Experimental Set-up 195 -- 5.4.2 Experiment 1: Activated Brain Region Selection using eLORETA 196 -- 5.4.3 Experiment 2: Detection of the ERP signals associated with the familiar andunfamiliar face discrimination 198 -- 5.4.4 Experiment 3: Performance Analysis of the Proposed CSP algorithm as a Feature extraction Technique 199 -- 5.4.5 Experiment 4: Performance Analysis of the Proposed LSTM based Classifier 201 -- 5.4.6 Experiment 5: Classifier Performance Analysis with varying EEG Time-Window Length 202 -- 5.4.7 Statistical Validation of the Proposed LSTM Classifier using McNamers' Test 203 -- 5.5 Conclusions 204 -- References 204 -- Chapter 6: Cognitive Load Assessment in Motor Learning Tasks by Near-Infrared Spectroscopy Using Type-2 Fuzzy Sets -- 6.1 Introduction 214 -- 6.2 Principles and Methodologies 216 -- 6.2.1 Normalization of Raw Data 217 -- 6.2.2 Pre-processing 218 -- 6.2.3 Feature Extraction 218 -- 6.2.4 Training Instance Generation for Offline Training 219 -- 6.2.5 Feature Selection using Evolutionary Algorithm 219 -- 6.2.6 Classifier Training and Testing 221 -- 6.3 Classifier Design 221 -- 6.3.1 Preliminaries of IT2FS and GT2FS 221 -- 6.3.2 IT2FS Induced Classifier Design 222 -- 6.3.3 GT2FS Induced Classifier Design 228 -- 6.4 Experiments and Results 230 -- 6.4.1 Experimental Set-up 230 -- 6.4.2 Participants 232 -- 6.4.3 Stimulus Presentation for Online Classification 232 -- 6.4.4 Experiment 1: Demonstration of decreasing Cognitive Load with increasing Learning Epochs for similar stimulus 233 -- 6 .4.5 Experiment 2: Automatic Extraction of Discriminating fNIRs features 234. 327 $a6.4.6 Experiment 3: Optimal Parameter Setting of Feature Selection and Classifier Units 235 -- 6.5 Biological Implications 237 -- 6.6 Performance Analysis 239 -- 6.6.1 Performance Analysis of the proposed IT2FS and GT2FS Classifier 239 -- 6.6.2 Statistical Validation of the Classifiers using McNamer;s Test 242 -- 6.7 Conclusion 243 -- References 243 -- Chapter 7: Conclusions and Future Directions of Research on BCI based Memory and Learning -- 7.1 Self-Review of the Works Undertaken in the Book 250 -- 7.2 Limitations of EEG BCI-Based Memory Experiments 252 -- 7.3 Further Scope of Future Research on Memory and Learning 253 -- References. 330 $a"This book models human memory from a cognitive standpoint by utilizing brain activations acquired from the cortex by electroencephalographic (EEG) and functional near-infrared-spectroscopic (f-NIRs) means. It begins with an overview of the early models of memory. The authors then propose a simplistic model of Working Memory (WM) built with fuzzy Hebbian learning. A second perspective of memory models is concerned with Short-Term Memory (STM)-modeling in the context of 2-dimensional object-shape reconstruction from visually examined memorized instances. A third model assesses the subjective motor learning skill in driving from erroneous motor actions. Other models introduce a novel strategy of designing a two-layered deep Long Short-Term Memory (LSTM) classifier network and also deal with cognitive load assessment in motor learning tasks associated with driving. The book ends with concluding remarks based on principles and experimental results acquired in previous chapters"--$cProvided by publisher 606 $aMemory 615 0$aMemory. 676 $a153.12 700 $aGhosh$b Lidia$01219995 702 $aKonar$b Amit 702 $aRakshit$b Pratyusha 801 0$bCaBNVSL 801 1$bCaBNVSL 801 2$bCaBNVSL 906 $aBOOK 912 $a9910555012703321 996 $aCognitive modeling of human memory and learning$92820885 997 $aUNINA LEADER 05263nam 2200625Ia 450 001 9910830495203321 005 20230617003304.0 010 $a1-280-27142-6 010 $a9786610271429 010 $a0-470-34579-9 010 $a0-470-86208-4 010 $a0-470-86207-6 035 $a(CKB)1000000000356056 035 $a(EBL)232710 035 $a(OCoLC)77503615 035 $a(SSID)ssj0000202743 035 $a(PQKBManifestationID)11201219 035 $a(PQKBTitleCode)TC0000202743 035 $a(PQKBWorkID)10255168 035 $a(PQKB)10734729 035 $a(MiAaPQ)EBC232710 035 $a(EXLCZ)991000000000356056 100 $a20040318d2004 uy 0 101 0 $aeng 135 $aur|n|---||||| 181 $ctxt 182 $cc 183 $acr 200 00$aMiddleware for communications$b[electronic resource] /$fedited by Qusay H. Mahmoud 210 $aChichester ;$aHoboken, NJ $cJohn Wiley & Sons$dc2004 215 $a1 online resource (523 p.) 300 $aDescription based upon print version of record. 311 $a0-470-86206-8 320 $aIncludes bibliographical references and index. 327 $aMiddleware for Communications; Contents; Preface; List of Contributors; Introduction; 1 Message-Oriented Middleware; 1.1 Introduction; 1.1.1 Interaction Models; 1.1.2 Synchronous Communication; 1.1.3 Asynchronous Communication; 1.1.4 Introduction to the Remote Procedure Call (RPC); 1.1.5 Introduction to Message-Oriented Middleware (MOM); 1.1.6 When to use MOM or RPC; 1.2 Message Queues; 1.3 Messaging Models; 1.3.1 Point-to-Point; 1.3.2 Publish/Subscribe; 1.3.3 Comparison of Messaging Models; 1.4 Common MOM Services; 1.4.1 Message Filtering; 1.4.2 Transactions 327 $a1.4.3 Guaranteed Message Delivery1.4.4 Message Formats; 1.4.5 Load Balancing; 1.4.6 Clustering; 1.5 Java Message Service; 1.5.1 Programming using the JMS API; 1.6 Service-Oriented Architectures; 1.6.1 XML; 1.6.2 Web Services; 1.6.3 MOM; 1.6.4 Developing Service-Oriented Architectures; 1.7 Summary; Bibliography; 2 Adaptive and Reflective Middleware; 2.1 Introduction; 2.1.1 Adaptive Middleware; 2.1.2 Reflective Middleware; 2.1.3 Are Adaptive and Reflective Techniques the Same?; 2.1.4 Triggers of Adaptive and Reflective Behavior; 2.2 Implementation Techniques; 2.2.1 Meta-Level Programming 327 $a2.2.2 Software Components and Frameworks2.2.3 Generative Programming; 2.3 Overview of Current Research; 2.3.1 Reflective and Adaptive Middleware Workshops; 2.3.2 Nonfunctional Properties; 2.3.3 Distribution Mechanism; 2.4 Future Research Directions; 2.4.1 Advances in Programming Techniques; 2.4.2 Open Research Issues; 2.4.3 Autonomic Computing; 2.5 Summary; Bibliography; 3 Transaction Middleware; 3.1 Introduction; 3.2 Transaction Processing Fundamentals; 3.2.1 ACID Transactions; 3.2.2 Distributed Transactions; 3.2.3 Common Extensions; 3.2.4 Programming Models for Transactions 327 $a3.3 Distributed Object Transactions3.3.1 Transaction Model; 3.3.2 Transaction APIs; 3.3.3 Container-Managed Transactions; 3.4 Messaging Transactions; 3.4.1 Messaging Models; 3.4.2 Programming Models; 3.4.3 Queued Transaction Processing; 3.5 Web Transactions; 3.5.1 Web Services Coordination and Transactions; 3.5.2 Programming model; 3.5.3 Web Services Messaging; 3.6 Advanced Transactions; 3.6.1 Long Running Unit of Work (LRUOW); 3.6.2 Conditional Messaging and D-Spheres; 3.6.3 Transactional Attitudes (TxA); 3.7 Conclusion; Bibliography; 4 Peer-to-Peer Middleware; 4.1 Introduction 327 $a4.1.1 Peer-to-Peer and Grids4.1.2 Lack of Peer-to-Peer Middleware; 4.1.3 Group Communication; 4.1.4 Challenges; 4.1.5 Chapter Outline; 4.2 JXTA; 4.2.1 Overview; 4.2.2 Resources and Advertisements; 4.2.3 Peer Groups; 4.2.4 Services and Modules; 4.2.5 Protocols; 4.2.6 Messages and Pipes; 4.2.7 Security; 4.2.8 Relay and Rendezvous Peers; 4.2.9 Group Communication; 4.2.10 Applications using JXTA; 4.2.11 Challenges; 4.2.12 Summary; 4.3 P2P Messaging System; 4.3.1 Self-Organizing Overlay Networks; 4.3.2 Failure Tolerance; 4.3.3 Implicit Dynamic Routing; 4.3.4 Quality-of-Service; 4.3.5 System Model 327 $a4.3.6 Network Abstraction Layer 330 $aA state-of-the-art guide to middleware technologies, and their pivotal role in communications networks. Middleware is about integration and interoperability of applications and services running on heterogeneous computing and communications devices. The services it provides - including identification, authentication, authorization, soft-switching, certification and security - are used in a vast range of global appliances and systems, from smart cards and wireless devices to mobile services and e-Commerce. Qusay H. Mahmoud has created an invaluable reference tool that explores the ori 606 $aTelecommunication systems$xData processing 606 $aMiddleware 615 0$aTelecommunication systems$xData processing. 615 0$aMiddleware. 676 $a005.3 676 $a005.7/13 701 $aMahmoud$b Qusay H.$f1971-$01622650 801 0$bMiAaPQ 801 1$bMiAaPQ 801 2$bMiAaPQ 906 $aBOOK 912 $a9910830495203321 996 $aMiddleware for communications$93956634 997 $aUNINA