Search results for: maze
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5

Search results for: maze

5 Recursive Path-finding in a Dynamic Maze with Modified Tremaux's Algorithm

Authors: Nien-Zheng Yew, Kung-Ming Tiong, Su-Ting Yong

Abstract:

Number Link is a Japanese logic puzzle where pairs of same numbers are connected using lines. Number Link can be regarded as a dynamic multiple travelers, multiple entries and exits maze, where the walls and passages are dynamically changing as the travelers move. In this paper, we apply the Tremaux’s algorithm to solve Number Link puzzles of size 8x8, 10x10 and 15x20. The algorithm works well and produces a solution for puzzles of size 8x8 and 10x10. However, solving a puzzle of size 15x20 requires high computer processing power and is time consuming.

Keywords: Number Link, maze, puzzle, Tremaux’s algorithm

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3978
4 Anxiolytic-like Effects of Dichloromethane Extracts of Valerian (DEV) in Adult Male Wistar Rats

Authors: Hadjikhani R.

Abstract:

Anxiety is a common disorder that attacks many people in society and often accompanied by physiological sensations such as tachycardia, chest pain, shortness of breath, insensitivity and etc. The purpose of this study is to characterize the putative anxiolytic-like effects of DEV (dichloromethane extracts of valerian) using the elevated plus maze (EPM) in rats. DEV was dissolved in DMSO and orally administered at different doses to adult male wistar rats, 0.5, 1.5 and 3 hours before behavioral evaluation in an EPM respectively. Control rats were treated with an equal volume of DMSO. Single treatment of DEV (at 0.1,0.2. 0.3, and 0.4 g/kg) significantly increased time-spent and arm entries into open arms of EPM versus control groups (p<0.05).However, no changes in the locomotor activity ccured. This result suggests that DEV might prove to be an effective anxiolytic agent.

Keywords: Anxiety, Dichloromethane extracts, Valerian, Rat

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1470
3 Capturing an Unknown Moving Target in Unknown Territory using Vision and Coordination

Authors: Kiran Ijaz, Umar Manzoor, Arshad Ali Shahid

Abstract:

In this paper we present an extension to Vision Based LRTA* (VLRTA*) known as Vision Based Moving Target Search (VMTS) for capturing unknown moving target in unknown territory with randomly generated obstacles. Target position is unknown to the agents and they cannot predict its position using any probability method. Agents have omni directional vision but can see in one direction at some point in time. Agent-s vision will be blocked by the obstacles in the search space so agent can not see through the obstacles. Proposed algorithm is evaluated on large number of scenarios. Scenarios include grids of sizes from 10x10 to 100x100. Grids had obstacles randomly placed, occupying 0% to 50%, in increments of 10%, of the search space. Experiments used 2 to 9 agents for each randomly generated maze with same obstacle ratio. Observed results suggests that VMTS is effective in locate target time, solution quality and virtual target. In addition, VMTS becomes more efficient if the number of agents is increased with proportion to obstacle ratio.

Keywords: Vision, MTS, Unknown Target, Coordination, VMTS, Multi-Agent.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1420
2 Robot Navigation and Localization Based on the Rat’s Brain Signals

Authors: Endri Rama, Genci Capi, Shigenori Kawahara

Abstract:

The mobile robot ability to navigate autonomously in its environment is very important. Even though the advances in technology, robot self-localization and goal directed navigation in complex environments are still challenging tasks. In this article, we propose a novel method for robot navigation based on rat’s brain signals (Local Field Potentials). It has been well known that rats accurately and rapidly navigate in a complex space by localizing themselves in reference to the surrounding environmental cues. As the first step to incorporate the rat’s navigation strategy into the robot control, we analyzed the rats’ strategies while it navigates in a multiple Y-maze, and recorded Local Field Potentials (LFPs) simultaneously from three brain regions. Next, we processed the LFPs, and the extracted features were used as an input in the artificial neural network to predict the rat’s next location, especially in the decision-making moment, in Y-junctions. We developed an algorithm by which the robot learned to imitate the rat’s decision-making by mapping the rat’s brain signals into its own actions. Finally, the robot learned to integrate the internal states as well as external sensors in order to localize and navigate in the complex environment.

Keywords: Brain machine interface, decision-making, local field potentials, mobile robot, navigation, neural network, rat, signal processing.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1421
1 A Probabilistic Reinforcement-Based Approach to Conceptualization

Authors: Hadi Firouzi, Majid Nili Ahmadabadi, Babak N. Araabi

Abstract:

Conceptualization strengthens intelligent systems in generalization skill, effective knowledge representation, real-time inference, and managing uncertain and indefinite situations in addition to facilitating knowledge communication for learning agents situated in real world. Concept learning introduces a way of abstraction by which the continuous state is formed as entities called concepts which are connected to the action space and thus, they illustrate somehow the complex action space. Of computational concept learning approaches, action-based conceptualization is favored because of its simplicity and mirror neuron foundations in neuroscience. In this paper, a new biologically inspired concept learning approach based on the probabilistic framework is proposed. This approach exploits and extends the mirror neuron-s role in conceptualization for a reinforcement learning agent in nondeterministic environments. In the proposed method, instead of building a huge numerical knowledge, the concepts are learnt gradually from rewards through interaction with the environment. Moreover the probabilistic formation of the concepts is employed to deal with uncertain and dynamic nature of real problems in addition to the ability of generalization. These characteristics as a whole distinguish the proposed learning algorithm from both a pure classification algorithm and typical reinforcement learning. Simulation results show advantages of the proposed framework in terms of convergence speed as well as generalization and asymptotic behavior because of utilizing both success and failures attempts through received rewards. Experimental results, on the other hand, show the applicability and effectiveness of the proposed method in continuous and noisy environments for a real robotic task such as maze as well as the benefits of implementing an incremental learning scenario in artificial agents.

Keywords: Concept learning, probabilistic decision making, reinforcement learning.

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1482