Open Access. Powered by Scholars. Published by Universities.®

Physical Sciences and Mathematics Commons

Open Access. Powered by Scholars. Published by Universities.®

Articles 1 - 6 of 6

Full-Text Articles in Physical Sciences and Mathematics

Planning With Ifalcon: Towards A Neural-Network-Based Bdi Agent Architecture, Budhitama Subagdja, Ah-Hwee Tan Dec 2008

Planning With Ifalcon: Towards A Neural-Network-Based Bdi Agent Architecture, Budhitama Subagdja, Ah-Hwee Tan

Research Collection School Of Computing and Information Systems

This paper presents iFALCON, a model of BDI (beliefdesire-intention) agents that is fully realized as a selforganizing neural network architecture. Based on multichannel network model called fusion ART, iFALCON is developed to bridge the gap between a self-organizing neural network that autonomously adapts its knowledge and the BDI agent model that follows explicit descriptions. Novel techniques called gradient encoding are introduced for representing sequences and hierarchical structures to realize plans and the intention structure. This paper shows that a simplified plan representation can be encoded as weighted connections in the neural network through a process of supervised learning. A case …


A Neural Network Model For A Hierarchical Spatio-Temporal Memory, Kiruthika Ramanathan, Luping Shi, Jianming Li, Kian Guan Lim, Zhi Ping Ang, Chong Chong Tow Nov 2008

A Neural Network Model For A Hierarchical Spatio-Temporal Memory, Kiruthika Ramanathan, Luping Shi, Jianming Li, Kian Guan Lim, Zhi Ping Ang, Chong Chong Tow

Research Collection School Of Computing and Information Systems

The architecture of the human cortex is uniform and hierarchical in nature. In this paper, we build upon works on hierarchical classification systems that model the cortex to develop a neural network representation for a hierarchical spatio-temporal memory (HST-M) system. The system implements spatial and temporal processing using neural network architectures. We have tested the algorithms developed against both the MLP and the Hierarchical Temporal Memory algorithms. Our results show definite improvement over MLP and are comparable to the performance of HTM.


Cascade Rsvm In Peer-To-Peer Network, Hock Hee Ang, Vivekanand Gopalkrishnan, Steven C. H. Hoi, Wee Keong Ng Sep 2008

Cascade Rsvm In Peer-To-Peer Network, Hock Hee Ang, Vivekanand Gopalkrishnan, Steven C. H. Hoi, Wee Keong Ng

Research Collection School Of Computing and Information Systems

The goal of distributed learning in P2P networks is to achieve results as close as possible to those from centralized approaches. Learning models of classification in a P2P network faces several challenges like scalability, peer dynamism, asynchronism and data privacy preservation. In this paper, we study the feasibility of building SVM classifiers in a P2P network. We show how cascading SVM can be mapped to a P2P network of data propagation. Our proposed P2P SVM provides a method for constructing classifiers in P2P networks with classification accuracy comparable to centralized classifiers and better than other distributed classifiers. The proposed algorithm …


Self-Organizing Neural Models Integrating Rules And Reinforcement Learning, Teck-Hou Teng, Zhong-Ming Tan, Ah-Hwee Tan Jun 2008

Self-Organizing Neural Models Integrating Rules And Reinforcement Learning, Teck-Hou Teng, Zhong-Ming Tan, Ah-Hwee Tan

Research Collection School Of Computing and Information Systems

Traditional approaches to integrating knowledge into neural network are concerned mainly about supervised learning. This paper presents how a family of self-organizing neural models known as fusion architecture for learning, cognition and navigation (FALCON) can incorporate a priori knowledge and perform knowledge refinement and expansion through reinforcement learning. Symbolic rules are formulated based on pre-existing know-how and inserted into FALCON as a priori knowledge. The availability of knowledge enables FALCON to start performing earlier in the initial learning trials. Through a temporal-difference (TD) learning method, the inserted rules can be refined and expanded according to the evaluative feedback signals received …


Traceable And Retrievable Identity-Based Encryption, Man Ho Au, Qiong Huang, Joseph K. Liu, Willy Susilo, Duncan S. Wong, Guomin Yang Jun 2008

Traceable And Retrievable Identity-Based Encryption, Man Ho Au, Qiong Huang, Joseph K. Liu, Willy Susilo, Duncan S. Wong, Guomin Yang

Research Collection School Of Computing and Information Systems

Very recently, the concept of Traceable Identity-based Encryption (IBE) scheme (or Accountable Authority Identity based Encryption scheme) was introduced in Crypto 2007. This concept enables some mechanisms to reduce the trust of a private key generator (PKG) in an IBE system. The aim of this paper is threefold. First, we discuss some subtleties in the first traceable IBE scheme in the Crypto 2007 paper. Second, we present an extension to this work by having the PKG’s master secret key retrieved automatically if more than one user secret key are released. This way, the user can produce a concrete proof of …


Integrating Temporal Difference Methods And Self‐Organizing Neural Networks For Reinforcement Learning With Delayed Evaluative Feedback, Ah-Hwee Tan, Ning Lu, Dan Xiao Feb 2008

Integrating Temporal Difference Methods And Self‐Organizing Neural Networks For Reinforcement Learning With Delayed Evaluative Feedback, Ah-Hwee Tan, Ning Lu, Dan Xiao

Research Collection School Of Computing and Information Systems

This paper presents a neural architecture for learning category nodes encoding mappings across multimodal patterns involving sensory inputs, actions, and rewards. By integrating adaptive resonance theory (ART) and temporal difference (TD) methods, the proposed neural model, called TD fusion architecture for learning, cognition, and navigation (TD-FALCON), enables an autonomous agent to adapt and function in a dynamic environment with immediate as well as delayed evaluative feedback (reinforcement) signals. TD-FALCON learns the value functions of the state-action space estimated through on-policy and off-policy TD learning methods, specifically state-action-reward-state-action (SARSA) and Q-learning. The learned value functions are then used to determine the …