SlideShare uma empresa Scribd logo
1 de 5
An Incremental Machine Learning Mechanism Applied to
                 Robot Navigation
            Nawwaf N Kharma, Majd Alwan, and Peter Y K Cheung
   Department of Electrical Engineering, Imperial College of Science Technology and Medicine,
                  London SW7 2BT U.K.           Fax: + 44 171 5814419
Abstract
In this paper we apply an incremental machine learning algorithm to the problem of robot navigation. The
learning algorithm is applied to a simple robot simulation to automatically induce a list of declarative rules. The
rules are pruned in order to remove the rules that are operationally useless. The final set is initially used to
control the robot navigating an obstacle-free path planned in a polygonal environment with satisfactory results.
Crisp conditions used in the rules are then replaced by fuzzy conditions fashioned by a human expert. The new
set of rules are shown to produce better results.
Keywords
incremental machine learning, tripartite rules, schema, robot navigation.
1. Introduction                                               to make it more suitable for sensorimotor robotic
                                                              applications. Our learning mechanism:
Both Classifier Systems and Q-Learning techniques
[1] have two major common deficiencies. These are:            * Aims at the automatic induction of a list of
                                                              declarative rules that describe the interaction between
A. The Two-Part Rule Form                                     an agent (e.g. robot) and its environment.
Production Rules (PR) (both classifiers and Q-                * Is simplified and made practically useful for real-
Learning rules are PR) have two parts only: A left side       time applications.
representing the conditions that have to be met before
the right side, the action, is taken. Thus PR in this         * Is amended to take into account the richness and the
form are situation-action rules. The information in           inherent uncertainties of the real world.
schemas (see 2.2) can be coded in a two-part PR               This paper has four main sections, the first represents
syntax. However, for many reasons this is not suitable.       the basic assumptions and terms that are needed to
1. There is evidence from animal ethology [2,3]               make the algorithm work. The second section
indicating that animals learn an action result                describes the main algorithm itself, the third outlines
association, and that this association, as a unit, is then    the specific problem and experiments carried out, and
linked with the context.                                      the fourth section shows and discusses the results
                                                              obtained.
2. The number of rules in PR systems is determined
by the number of combinations of contexts, actions,           2. Assumptions and Terms
and results that could make up a rule. This could             2.1 Basic Assumptions
result in a very large number of rules. In contrast,
schemas used in this paper are built incrementally and        The following are the basic assumptions made, in
hence require less memory.                                    order for the main algorithm to work in line with
                                                              expectations:
B. Implicit Representation of Result Values                   * All learned information may be put in the form of a
1. Q-learning as well as Classifier Systems assign a          list of declarative rules.
strength to each rule that implicitly expresses the           * The nature of the environment is static. The set of
operational value the rule. This contrasts to schemas         laws that govern the environment do not change over
that have explicit declarative result components.             time.
                                                              * There are no hidden states. The relevant aspects of
2. The rule selection mechanism in PR systems                 the environment are all detectable through the robot’s
chooses high strength rules, or rules that are in the         sensors.
vicinity of high strength ones. This means that               * Crisp conditions are initially sufficient for learning.
learning only takes place along the fringe of the state-      * Disjunctions of conjunctions of conditions are
space that has already been connected to a goal. While        enough to characterise any state.
animats should be allowed to seek knowledge that              * Temporal Credit Allocation problem [5] may be
may not have immediate use for the goal at hand.              overlooked.
The Learning Mechanism on the other hand is a                 * Actions are taken in serial. They are finite in
system that learns incrementally (i.e. every rule is          duration, and do not destabilise the system.
built in a number of steps) using explicit units of           * Relevant results can be pre-defined in terms of a
representation (schemas). The algortihm aims to               combination of conditions.
enable robots to acquire and use sensorimotor                 * There are any number of agents in the environment.
knowledge autonomously, without a priori knowledge            Any one of them may be monitored by the learning
of the skills concerned. This algorithm is based on the       algorithm.
Schema Mechanism developed by Drescher [4]. It was            2.2 Definiton of Terms
altered and amended significantly in three main ways
                                                             • Schema and Schema Space
result follow more reliably. Reliability of a schema is
                  context          result                      measured by the ratio of: the number of times that its
                          action
                    extended structures
                                                               action is executed, in the right context, and leads to
                                                               the fulfilment of its result, to the total number of times
                         schemai                               that its action is executed in the right context.
                     Fig. 1 A schema.                         • Configuration Parameters and others
  The main structure of the learning mechanism is the          - Result spin-off: a new schema made out of a copy of
  Schema (or rule) Space. The Schema Space is the              a previous one, by adding a condition to the result
  collection of all schemas. At any time the Schema            side.
  Space of the robot represents all its knowledge. The
  job of the learning algorithm is simply to create,           - Context spin-off: a new schema made out of a copy
  modify, delete, and possibly link schemas.                   of a previous one, by adding a condition to the context
                                                               side.
  A schema representation is made of two main
  structures: a Main Body (which comprises of a                - θ1: the relevance threshold, for producing result
  context, action and result), and the extended                spin-offs.
  structures. (see Fig. 1.) The main body is a tripartite      - N1: the total number of experiments that need to be
  rule representing a counterfactual assertion. The            taken before a result spin-off is allowed.
  extended structures keep information that is mainly
  used for creating (or spinning-off) new schemas.             - θ2: the reliability threshold, used for producing
                                                               context spin-offs.
  A schema has both declarative and procedural aspects.
                                                               - N2: The number of activations that a result spin-off
  Declaratively, it is a unit of information about an          schema needs to go through before it is allowed to
  interaction between the robot and the environment.           produce a context spin-off.
  Procedurally, a schema represents a possible action to
  be taken at situations when its context is fulfilled and     3. The Main Algorithm
  its result is desired.                                       The learning mechanism is best described by
                                                               explaining the main algorithm that it embodies. This
  The components of a schema are:
                                                               main algorithm goes through the following main
• Main Body:                                                   steps:
  - Conditions: A condition may be viewed as a                 1. Randomly select an action and execute it.
  function representing the degree of membership (or           2. Use the data collected before, during and after
  D.O.M.) of a sensor's output in a set representing that      taking the action of the schema in its context, to
  condition. In a crisp DOM case, a condition can either       update the two sets of correlation statistics
  be true or false.                                            3. Based on the statistics in step 2, the rule base may
                                                               be updated as detailed in the algorithmic notation.
  - Context, Result and Action: A Context (and                 4. Repeat steps 1 to 3 above until the predetermined
  similarly a Result) is a conjunction of one or more          number of experiments is met.
  conditions (and their negations.) A Result could be
  either predefined or created at run-time. Contexts of        The two phases of rule base update is best described in
  reliable schemas are automatically added, at run-time,       the following algorithmic notation:
  to the set of results. An Action represents a command                 If       ( no of experiments > N1
  to an effector to take an action. If an Action is taken                        AND PTC/NTC(Resulti) >= θ1
  then it's command is executed.                                                 AND Resulti not used before)
• Extended Structures                                                   then     Result spin-off
  Each schema has extended structures that contain two         When update is completed and once the PSC and NSC
  main sets of correlation statistics. These statistics are    are known context spin-off takes place according to:
  necessary for the development of schemas. The first                   If       (no. of experiments > N2
  set contains the Positive Transition Correlation (PTC)                         AND PSC/NSC(Conditioni) >= θ2
  and the Negative Transition Correlation (NTC) are                              AND Conditioni not used before)
  used to find relevant results of an action. A relevant                then     Context spin-off
  result of an action is a result that has empirically
  shown that it follows the execution of that action           4. Problem and Experiments
  significantly more often than other actions from the         The learning algorithm is now applied to the problem
  robot’s repertoire of actions. The PTC discovers             of robot navigation. The goal of this application is to:
  positive results while the NTC discovers negative
  ones. The second set of statistics contains the Positive     * Show that the algorithm is capable to deduce a list
  Success Correlation (PSC) and the Negative Success           of rules that is capable (if properly pruned) of
  Correlation (NSC). PSC is used to find conditions that       controlling the navigational behaviour of the robot
  when included in the context of a schema, will make          navigating an obstacle-free path planned in a given
  its result follow more reliably than before adding           environment.
  these conditions. NSC has the same function as PSC           * Investigate the results of fuzzifying the
  except that it is used to find conditions that need to be    context/result conditions on the execution of the
  excluded from the context of a schema to make its            deduced rule base.
4.1 The Robot Simulation and the Task to Learn              IF right_big ^ left_slight       THEN right_small
The robot has a cylindrical body, a differential drive      IF right_small ^ left_slight     THEN centre
with two independently driven motorised wheels that         IF centre ^ left_slight          THEN left_small
perform both the diving and steering. Four castors to
support the mobile base on the flour (see Fig. 2.)          IF left_small ^ left_slight      THEN left_big
                    C              C                      They were found with different reliability values
                                                          depending on the specific series of experiments taken.
             DW L                      DW R               The rules produced by the learning algorithm are then
                                                          pruned using the criteria of:-
                                                          1. Relevance to the goal (heading towards the goal),
                        C      C
                                                          2. High reliability.
   Fig. 2 A schematic of the Mobile Robot’s Base.
                                                          The above rules become:
Steering results from driving the left and right wheels
at different speeds. This arrangement enables the robot     IF right_big ^ left_slight       THEN right_small
to turn around its centre.                                  IF right_small ^ left_slight     THEN centre
The robot is equipped with an on-board electronic         With respect to rule block 1, the final list becomes
compass, and odometry for localisation.                   (put in operational form):
The robot requires two commands, linear speed and           IF DirDif: right_big      THEN DirOut: left_far
change of direction. These are separated into
individual rotational speed commands for the two            IF right_big              THEN left_slight
driving motors, which are put in a velocity closed-         IF right_small            THEN left_far
loop control. The global position control loop is
                                                            IF right_small            THEN left_slight
closed by the feedback coming from the localisation
system.                                                     IF centre                 THEN straight
The task we want to learn is navigating our robot on a      IF left_small             THEN right_slight
path consisting of straight line segments. The learnt       IF left_small             THEN right_far
navigation rulebase should be able to control the robot
to traverse the planned path smoothly.                      IF left_big               THEN right_slight
A simulation of the kinematics and dynamics of the          IF left_big               THEN right_far
described mobile robot base was used for testing the      For the second block of rules, those that are concerned
learnt control rules. The robot simulation links to       with the control of the linear velocity of the robot
FuzzyTECH 3.1 development environment [6], where          when heading towards a goal, a number of constraints
the rules and the input/output membership functions       is placed on the learning algorithm:-
(including crisp ones) can be graphically edited.
                                                          1. Due to inertia, the robot is prevented from taking an
4.2 Experimental set-up for learning                      experiment in which the speed changes suddenly from
The learning algorithm goes through two runs. One to      slow to high or high to zero. Speed can only change
discover the block of rules that are relevant to the      gradually. This corresponds to real robots with
orientation control. The second block contains rules      dynamics, as opposed to mere kinematic simulations.
that control the linear velocity of the robot.            2. We prune the first list of rules according to a
                                                          different criteria (from the case with the orientation
The learning algorithm is configured as follows: θ1:=     block). This criteria is:
2, N1:= the total number of experiments taken, θ2:= 1,
N2:= 3. Negative spin-off mechanisms are disabled.                 A. Highest reliability.
                                                                   B. Maximum distance traversal at each step.
The sets of conditions and actions used are:                       C. Zero speed at the goal.
DirDif={right_big, right_small, centre, left_small,       This gives us the following list of rules:
left_big}, Dist={very_near, near, medium, far},
SpIn={zero,       slow,        medium,         high},       IF Dist: X ^ SpIn: zero      THEN SpOut: slow
DirOut={left_far, left_slight, straight, right_slight,      IF medium ^ medium           THEN medium
right_far}, SpOut={ zero, slow, medium, high }.
                                                            IF far ^ medium              THEN high
5. Results
                                                            IF far ^ slow                THEN medium
5.1 Learning Algorithm Results
                                                            IF far ^ high                THEN high
A series of experiments are fed to the learning
algorithm. These experiments were chosen such that          IF medium ^ high             THEN medium
they cover, on a uniformly random basis, the context        IF medium ^ slow             THEN medium
space of the actions concerned. The learning
                                                            IF near ^ slow               THEN slow
algorithm is run and a series of rules are produced. If
the direction control action left_slight is taken as an     IF very_near ^ slow          THEN zero
example we will find the          following rules are
produced:
Since the two blocks of rules are learned separately,    However, when appropriate fuzzy membership
separation is enforced in action. This is done via       functions replace the crisp ones, the performance of
adding another block of rules which makes sure that      the learnt navigation rules significantly improves, as
the speed rules are only active when the robot is        Fig. 4 shows. This is because when the robot becomes
heading towards the goal. This special block is:         closer to the direction of the goal, the final output of
                                                         the orientation control rules is significantly reduced
  IF DirOut: right_far     THEN SpOut: zero
                                                         according to the degree of fulfilment.
  IF left_far              THEN zero
                                                         6. Conclusions and Recommendations
  IF right_slight          THEN zero
                                                         The learning algorithm succeeded in finding the
  IF left_slight           THEN zero                     declarative rules that represent, in their totality, the
This means that at execution the robot should first      interaction between the robot and the environment.
execute the first set making sure that robot is in the   Many of these rules were operationally useless, and
right direction and then the second block starts         had to be pruned (according to the criteria mentioned
executing.                                               previously). Once pruned, the resulting rules (both in
                                                         crisp and fuzzy forms) were effective in controlling
5.2 Simulation Results                                   the robot in navigation.
                                                         We have shown that the performance of the learnt
                                                         schemas improves as the context conditions are
                                                         fuzzified. Hence, our future work would be making
                                                         the learning schema mechanism a fuzzy one, which
                                                         would be more general and capable of learning tasks
                                                         in the continuous real world. Our learning mechanism,
                                                         presented in this paper, readily allows this extension.
                                                         7. References
                                                         [1] Dorigo M. Et al. (1994) "A comparison of Q-learning and
                                                         classifier systems." In From animals to animats 3, edited by D.Cliff et
                                                         al. MIT Press, Cambridge, MA.
                                                         [2] Rescorla R. (1990) "Evidence for an association between the
                                                         disciminative stimulus and the response-outcome association in
                                                         instrumental learning." Journal of experimental psychology: animal
                                                         behavior process, 16, 326-334.
      Fig. 3 Navigation using Crisp Conditions.          [3] Roitblat H.(1994) "Mechanism and process in animal behavior:
                                                         models of animals, animals as models." In From animals to animats
Fig. 3 shows the robot navigating a planned path using   3, edited by D.Cliff et al. MIT Press, Cambridge, MA.
the learnt rules with crisp membership functions for     [4] Drescher G. (1990) "Made-up minds: a constructivist approach to
the context conditions. It is clear that the robot’s     artificial intelligence." MIT Press, Cambridge, MA.
centre moved off the straight path segments, due to      [5] Holland J H. (1992) “Adaptation in Natural and Artificial
non-overlapping between the straight and the             Systems.” MIT Press, Cambridge, MA.
contiguous membership functions, and to its width.       [6] FuzzyTECH 3.1 Software Manuals.
This is unsuitable in cluttered environments (e.g. a
narrow corridor). Had the straight membership
function been narrower the robot would have swung
right and left of the path in a zigzag, due to the
activation of exactly the same rules regardless of the
required amount of direction change when change of
direction is required.




     Fig. 4 Navigation using Fuzzy Conditions.

Mais conteúdo relacionado

Mais procurados

A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...JPINFOTECH JAYAPRAKASH
 
A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...IEEEFINALYEARPROJECTS
 
a deep reinforced model for abstractive summarization
a deep reinforced model for abstractive summarizationa deep reinforced model for abstractive summarization
a deep reinforced model for abstractive summarizationJEE HYUN PARK
 
Reinforcement learning-ebook-part1
Reinforcement learning-ebook-part1Reinforcement learning-ebook-part1
Reinforcement learning-ebook-part1Rajmeet Singh
 
Decentralized Data Fusion Algorithm using Factor Analysis Model
Decentralized Data Fusion Algorithm using Factor Analysis ModelDecentralized Data Fusion Algorithm using Factor Analysis Model
Decentralized Data Fusion Algorithm using Factor Analysis ModelSayed Abulhasan Quadri
 
Iaetsd an efficient and large data base using subset selection algorithm
Iaetsd an efficient and large data base using subset selection algorithmIaetsd an efficient and large data base using subset selection algorithm
Iaetsd an efficient and large data base using subset selection algorithmIaetsd Iaetsd
 
Intelligent Controller Design for a Chemical Process
Intelligent Controller Design for a Chemical ProcessIntelligent Controller Design for a Chemical Process
Intelligent Controller Design for a Chemical ProcessCSCJournals
 
Draft comparison of electronic reliability prediction methodologies
Draft comparison of electronic reliability prediction methodologiesDraft comparison of electronic reliability prediction methodologies
Draft comparison of electronic reliability prediction methodologiesAccendo Reliability
 
Main assignment 1
Main assignment 1Main assignment 1
Main assignment 1Tinu P Saju
 
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...inventionjournals
 
Rule Evaluation on a Motorola SIMD
Rule Evaluation on a Motorola SIMDRule Evaluation on a Motorola SIMD
Rule Evaluation on a Motorola SIMDMeltin Bell
 
An Algorithm of Policy Gradient Reinforcement Learning with a Fuzzy Controlle...
An Algorithm of Policy Gradient Reinforcement Learning with a Fuzzy Controlle...An Algorithm of Policy Gradient Reinforcement Learning with a Fuzzy Controlle...
An Algorithm of Policy Gradient Reinforcement Learning with a Fuzzy Controlle...Waqas Tariq
 

Mais procurados (15)

Differential evolution optimization technique
Differential evolution optimization techniqueDifferential evolution optimization technique
Differential evolution optimization technique
 
A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...
 
A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...
 
a deep reinforced model for abstractive summarization
a deep reinforced model for abstractive summarizationa deep reinforced model for abstractive summarization
a deep reinforced model for abstractive summarization
 
Optimization Using Evolutionary Computing Techniques
Optimization Using Evolutionary Computing Techniques Optimization Using Evolutionary Computing Techniques
Optimization Using Evolutionary Computing Techniques
 
Fuzzy model reference learning control (1)
Fuzzy model reference learning control (1)Fuzzy model reference learning control (1)
Fuzzy model reference learning control (1)
 
Reinforcement learning-ebook-part1
Reinforcement learning-ebook-part1Reinforcement learning-ebook-part1
Reinforcement learning-ebook-part1
 
Decentralized Data Fusion Algorithm using Factor Analysis Model
Decentralized Data Fusion Algorithm using Factor Analysis ModelDecentralized Data Fusion Algorithm using Factor Analysis Model
Decentralized Data Fusion Algorithm using Factor Analysis Model
 
Iaetsd an efficient and large data base using subset selection algorithm
Iaetsd an efficient and large data base using subset selection algorithmIaetsd an efficient and large data base using subset selection algorithm
Iaetsd an efficient and large data base using subset selection algorithm
 
Intelligent Controller Design for a Chemical Process
Intelligent Controller Design for a Chemical ProcessIntelligent Controller Design for a Chemical Process
Intelligent Controller Design for a Chemical Process
 
Draft comparison of electronic reliability prediction methodologies
Draft comparison of electronic reliability prediction methodologiesDraft comparison of electronic reliability prediction methodologies
Draft comparison of electronic reliability prediction methodologies
 
Main assignment 1
Main assignment 1Main assignment 1
Main assignment 1
 
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
 
Rule Evaluation on a Motorola SIMD
Rule Evaluation on a Motorola SIMDRule Evaluation on a Motorola SIMD
Rule Evaluation on a Motorola SIMD
 
An Algorithm of Policy Gradient Reinforcement Learning with a Fuzzy Controlle...
An Algorithm of Policy Gradient Reinforcement Learning with a Fuzzy Controlle...An Algorithm of Policy Gradient Reinforcement Learning with a Fuzzy Controlle...
An Algorithm of Policy Gradient Reinforcement Learning with a Fuzzy Controlle...
 

Destaque

【作业】《传播学理论》
【作业】《传播学理论》【作业】《传播学理论》
【作业】《传播学理论》qian jiaqing
 
Timeline: The Life of Michael Jackson
Timeline: The Life of Michael JacksonTimeline: The Life of Michael Jackson
Timeline: The Life of Michael Jacksonbutest
 
LESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALLESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALbutest
 
1. MPEG I.B.P frame之不同
1. MPEG I.B.P frame之不同1. MPEG I.B.P frame之不同
1. MPEG I.B.P frame之不同butest
 
Tearn Up pitch deck.pdf
Tearn Up pitch deck.pdfTearn Up pitch deck.pdf
Tearn Up pitch deck.pdfasenju
 

Destaque (7)

WORD
WORDWORD
WORD
 
New Media Language
New Media LanguageNew Media Language
New Media Language
 
【作业】《传播学理论》
【作业】《传播学理论》【作业】《传播学理论》
【作业】《传播学理论》
 
Timeline: The Life of Michael Jackson
Timeline: The Life of Michael JacksonTimeline: The Life of Michael Jackson
Timeline: The Life of Michael Jackson
 
LESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALLESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIAL
 
1. MPEG I.B.P frame之不同
1. MPEG I.B.P frame之不同1. MPEG I.B.P frame之不同
1. MPEG I.B.P frame之不同
 
Tearn Up pitch deck.pdf
Tearn Up pitch deck.pdfTearn Up pitch deck.pdf
Tearn Up pitch deck.pdf
 

Semelhante a .doc

SELF LEARNING REAL TIME EXPERT SYSTEM
SELF LEARNING REAL TIME EXPERT SYSTEMSELF LEARNING REAL TIME EXPERT SYSTEM
SELF LEARNING REAL TIME EXPERT SYSTEMcscpconf
 
Self learning real time expert system
Self learning real time expert systemSelf learning real time expert system
Self learning real time expert systemijscai
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural networkNagarajan
 
Survey on Artificial Neural Network Learning Technique Algorithms
Survey on Artificial Neural Network Learning Technique AlgorithmsSurvey on Artificial Neural Network Learning Technique Algorithms
Survey on Artificial Neural Network Learning Technique AlgorithmsIRJET Journal
 
Literature Survey
Literature SurveyLiterature Survey
Literature Surveybutest
 
17 2 expert systems
17 2 expert systems17 2 expert systems
17 2 expert systemsTianlu Wang
 
Model-based Investigation of the Effect of Tuning Parameters o.docx
Model-based Investigation of the Effect of Tuning Parameters o.docxModel-based Investigation of the Effect of Tuning Parameters o.docx
Model-based Investigation of the Effect of Tuning Parameters o.docxraju957290
 
Benchmark methods to analyze embedded processors and systems
Benchmark methods to analyze embedded processors and systemsBenchmark methods to analyze embedded processors and systems
Benchmark methods to analyze embedded processors and systemsXMOS
 
Discrete event systems comprise of discrete state spaces and event
Discrete event systems comprise of discrete state spaces and eventDiscrete event systems comprise of discrete state spaces and event
Discrete event systems comprise of discrete state spaces and eventNitish Nagar
 
Behavior modeling of soft real time system using stereotyped extension mechan...
Behavior modeling of soft real time system using stereotyped extension mechan...Behavior modeling of soft real time system using stereotyped extension mechan...
Behavior modeling of soft real time system using stereotyped extension mechan...eSAT Publishing House
 
A boolean modeling for improving
A boolean modeling for improvingA boolean modeling for improving
A boolean modeling for improvingcsandit
 
Reinforcement Learning / E-Book / Part 1
Reinforcement Learning / E-Book / Part 1Reinforcement Learning / E-Book / Part 1
Reinforcement Learning / E-Book / Part 1Hitesh Mohapatra
 
Knowledge-Based Agent in Artificial intelligence.pptx
Knowledge-Based Agent in Artificial intelligence.pptxKnowledge-Based Agent in Artificial intelligence.pptx
Knowledge-Based Agent in Artificial intelligence.pptxsuchita74
 
Adapted Branch-and-Bound Algorithm Using SVM With Model Selection
Adapted Branch-and-Bound Algorithm Using SVM With Model SelectionAdapted Branch-and-Bound Algorithm Using SVM With Model Selection
Adapted Branch-and-Bound Algorithm Using SVM With Model SelectionIJECEIAES
 
Chapter 1 Data structure.pptx
Chapter 1 Data structure.pptxChapter 1 Data structure.pptx
Chapter 1 Data structure.pptxwondmhunegn
 
Rachit Mishra_stock prediction_report
Rachit Mishra_stock prediction_reportRachit Mishra_stock prediction_report
Rachit Mishra_stock prediction_reportRachit Mishra
 
New folderIMAG2318.jpgNew folderIMAG2319.jpgNew folder.docx
New folderIMAG2318.jpgNew folderIMAG2319.jpgNew folder.docxNew folderIMAG2318.jpgNew folderIMAG2319.jpgNew folder.docx
New folderIMAG2318.jpgNew folderIMAG2319.jpgNew folder.docxhenrymartin15260
 

Semelhante a .doc (20)

SELF LEARNING REAL TIME EXPERT SYSTEM
SELF LEARNING REAL TIME EXPERT SYSTEMSELF LEARNING REAL TIME EXPERT SYSTEM
SELF LEARNING REAL TIME EXPERT SYSTEM
 
Self learning real time expert system
Self learning real time expert systemSelf learning real time expert system
Self learning real time expert system
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural network
 
Survey on Artificial Neural Network Learning Technique Algorithms
Survey on Artificial Neural Network Learning Technique AlgorithmsSurvey on Artificial Neural Network Learning Technique Algorithms
Survey on Artificial Neural Network Learning Technique Algorithms
 
Literature Survey
Literature SurveyLiterature Survey
Literature Survey
 
17 2 expert systems
17 2 expert systems17 2 expert systems
17 2 expert systems
 
Model-based Investigation of the Effect of Tuning Parameters o.docx
Model-based Investigation of the Effect of Tuning Parameters o.docxModel-based Investigation of the Effect of Tuning Parameters o.docx
Model-based Investigation of the Effect of Tuning Parameters o.docx
 
Production System
Production SystemProduction System
Production System
 
Benchmark methods to analyze embedded processors and systems
Benchmark methods to analyze embedded processors and systemsBenchmark methods to analyze embedded processors and systems
Benchmark methods to analyze embedded processors and systems
 
Discrete event systems comprise of discrete state spaces and event
Discrete event systems comprise of discrete state spaces and eventDiscrete event systems comprise of discrete state spaces and event
Discrete event systems comprise of discrete state spaces and event
 
Behavior modeling of soft real time system using stereotyped extension mechan...
Behavior modeling of soft real time system using stereotyped extension mechan...Behavior modeling of soft real time system using stereotyped extension mechan...
Behavior modeling of soft real time system using stereotyped extension mechan...
 
A boolean modeling for improving
A boolean modeling for improvingA boolean modeling for improving
A boolean modeling for improving
 
MODELING & SIMULATION.docx
MODELING & SIMULATION.docxMODELING & SIMULATION.docx
MODELING & SIMULATION.docx
 
10.1.1.97.5586
10.1.1.97.558610.1.1.97.5586
10.1.1.97.5586
 
Reinforcement Learning / E-Book / Part 1
Reinforcement Learning / E-Book / Part 1Reinforcement Learning / E-Book / Part 1
Reinforcement Learning / E-Book / Part 1
 
Knowledge-Based Agent in Artificial intelligence.pptx
Knowledge-Based Agent in Artificial intelligence.pptxKnowledge-Based Agent in Artificial intelligence.pptx
Knowledge-Based Agent in Artificial intelligence.pptx
 
Adapted Branch-and-Bound Algorithm Using SVM With Model Selection
Adapted Branch-and-Bound Algorithm Using SVM With Model SelectionAdapted Branch-and-Bound Algorithm Using SVM With Model Selection
Adapted Branch-and-Bound Algorithm Using SVM With Model Selection
 
Chapter 1 Data structure.pptx
Chapter 1 Data structure.pptxChapter 1 Data structure.pptx
Chapter 1 Data structure.pptx
 
Rachit Mishra_stock prediction_report
Rachit Mishra_stock prediction_reportRachit Mishra_stock prediction_report
Rachit Mishra_stock prediction_report
 
New folderIMAG2318.jpgNew folderIMAG2319.jpgNew folder.docx
New folderIMAG2318.jpgNew folderIMAG2319.jpgNew folder.docxNew folderIMAG2318.jpgNew folderIMAG2319.jpgNew folder.docx
New folderIMAG2318.jpgNew folderIMAG2319.jpgNew folder.docx
 

Mais de butest

EL MODELO DE NEGOCIO DE YOUTUBE
EL MODELO DE NEGOCIO DE YOUTUBEEL MODELO DE NEGOCIO DE YOUTUBE
EL MODELO DE NEGOCIO DE YOUTUBEbutest
 
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...butest
 
LESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALLESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALbutest
 
Com 380, Summer II
Com 380, Summer IICom 380, Summer II
Com 380, Summer IIbutest
 
The MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
The MYnstrel Free Press Volume 2: Economic Struggles, Meet JazzThe MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
The MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazzbutest
 
MICHAEL JACKSON.doc
MICHAEL JACKSON.docMICHAEL JACKSON.doc
MICHAEL JACKSON.docbutest
 
Social Networks: Twitter Facebook SL - Slide 1
Social Networks: Twitter Facebook SL - Slide 1Social Networks: Twitter Facebook SL - Slide 1
Social Networks: Twitter Facebook SL - Slide 1butest
 
Facebook
Facebook Facebook
Facebook butest
 
Executive Summary Hare Chevrolet is a General Motors dealership ...
Executive Summary Hare Chevrolet is a General Motors dealership ...Executive Summary Hare Chevrolet is a General Motors dealership ...
Executive Summary Hare Chevrolet is a General Motors dealership ...butest
 
Welcome to the Dougherty County Public Library's Facebook and ...
Welcome to the Dougherty County Public Library's Facebook and ...Welcome to the Dougherty County Public Library's Facebook and ...
Welcome to the Dougherty County Public Library's Facebook and ...butest
 
NEWS ANNOUNCEMENT
NEWS ANNOUNCEMENTNEWS ANNOUNCEMENT
NEWS ANNOUNCEMENTbutest
 
C-2100 Ultra Zoom.doc
C-2100 Ultra Zoom.docC-2100 Ultra Zoom.doc
C-2100 Ultra Zoom.docbutest
 
MAC Printing on ITS Printers.doc.doc
MAC Printing on ITS Printers.doc.docMAC Printing on ITS Printers.doc.doc
MAC Printing on ITS Printers.doc.docbutest
 
Mac OS X Guide.doc
Mac OS X Guide.docMac OS X Guide.doc
Mac OS X Guide.docbutest
 
WEB DESIGN!
WEB DESIGN!WEB DESIGN!
WEB DESIGN!butest
 
Download
DownloadDownload
Downloadbutest
 
resume.doc
resume.docresume.doc
resume.docbutest
 
Download.doc.doc
Download.doc.docDownload.doc.doc
Download.doc.docbutest
 

Mais de butest (20)

EL MODELO DE NEGOCIO DE YOUTUBE
EL MODELO DE NEGOCIO DE YOUTUBEEL MODELO DE NEGOCIO DE YOUTUBE
EL MODELO DE NEGOCIO DE YOUTUBE
 
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
 
LESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALLESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIAL
 
Com 380, Summer II
Com 380, Summer IICom 380, Summer II
Com 380, Summer II
 
PPT
PPTPPT
PPT
 
The MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
The MYnstrel Free Press Volume 2: Economic Struggles, Meet JazzThe MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
The MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
 
MICHAEL JACKSON.doc
MICHAEL JACKSON.docMICHAEL JACKSON.doc
MICHAEL JACKSON.doc
 
Social Networks: Twitter Facebook SL - Slide 1
Social Networks: Twitter Facebook SL - Slide 1Social Networks: Twitter Facebook SL - Slide 1
Social Networks: Twitter Facebook SL - Slide 1
 
Facebook
Facebook Facebook
Facebook
 
Executive Summary Hare Chevrolet is a General Motors dealership ...
Executive Summary Hare Chevrolet is a General Motors dealership ...Executive Summary Hare Chevrolet is a General Motors dealership ...
Executive Summary Hare Chevrolet is a General Motors dealership ...
 
Welcome to the Dougherty County Public Library's Facebook and ...
Welcome to the Dougherty County Public Library's Facebook and ...Welcome to the Dougherty County Public Library's Facebook and ...
Welcome to the Dougherty County Public Library's Facebook and ...
 
NEWS ANNOUNCEMENT
NEWS ANNOUNCEMENTNEWS ANNOUNCEMENT
NEWS ANNOUNCEMENT
 
C-2100 Ultra Zoom.doc
C-2100 Ultra Zoom.docC-2100 Ultra Zoom.doc
C-2100 Ultra Zoom.doc
 
MAC Printing on ITS Printers.doc.doc
MAC Printing on ITS Printers.doc.docMAC Printing on ITS Printers.doc.doc
MAC Printing on ITS Printers.doc.doc
 
Mac OS X Guide.doc
Mac OS X Guide.docMac OS X Guide.doc
Mac OS X Guide.doc
 
hier
hierhier
hier
 
WEB DESIGN!
WEB DESIGN!WEB DESIGN!
WEB DESIGN!
 
Download
DownloadDownload
Download
 
resume.doc
resume.docresume.doc
resume.doc
 
Download.doc.doc
Download.doc.docDownload.doc.doc
Download.doc.doc
 

.doc

  • 1. An Incremental Machine Learning Mechanism Applied to Robot Navigation Nawwaf N Kharma, Majd Alwan, and Peter Y K Cheung Department of Electrical Engineering, Imperial College of Science Technology and Medicine, London SW7 2BT U.K. Fax: + 44 171 5814419
  • 2. Abstract In this paper we apply an incremental machine learning algorithm to the problem of robot navigation. The learning algorithm is applied to a simple robot simulation to automatically induce a list of declarative rules. The rules are pruned in order to remove the rules that are operationally useless. The final set is initially used to control the robot navigating an obstacle-free path planned in a polygonal environment with satisfactory results. Crisp conditions used in the rules are then replaced by fuzzy conditions fashioned by a human expert. The new set of rules are shown to produce better results. Keywords incremental machine learning, tripartite rules, schema, robot navigation. 1. Introduction to make it more suitable for sensorimotor robotic applications. Our learning mechanism: Both Classifier Systems and Q-Learning techniques [1] have two major common deficiencies. These are: * Aims at the automatic induction of a list of declarative rules that describe the interaction between A. The Two-Part Rule Form an agent (e.g. robot) and its environment. Production Rules (PR) (both classifiers and Q- * Is simplified and made practically useful for real- Learning rules are PR) have two parts only: A left side time applications. representing the conditions that have to be met before the right side, the action, is taken. Thus PR in this * Is amended to take into account the richness and the form are situation-action rules. The information in inherent uncertainties of the real world. schemas (see 2.2) can be coded in a two-part PR This paper has four main sections, the first represents syntax. However, for many reasons this is not suitable. the basic assumptions and terms that are needed to 1. There is evidence from animal ethology [2,3] make the algorithm work. The second section indicating that animals learn an action result describes the main algorithm itself, the third outlines association, and that this association, as a unit, is then the specific problem and experiments carried out, and linked with the context. the fourth section shows and discusses the results obtained. 2. The number of rules in PR systems is determined by the number of combinations of contexts, actions, 2. Assumptions and Terms and results that could make up a rule. This could 2.1 Basic Assumptions result in a very large number of rules. In contrast, schemas used in this paper are built incrementally and The following are the basic assumptions made, in hence require less memory. order for the main algorithm to work in line with expectations: B. Implicit Representation of Result Values * All learned information may be put in the form of a 1. Q-learning as well as Classifier Systems assign a list of declarative rules. strength to each rule that implicitly expresses the * The nature of the environment is static. The set of operational value the rule. This contrasts to schemas laws that govern the environment do not change over that have explicit declarative result components. time. * There are no hidden states. The relevant aspects of 2. The rule selection mechanism in PR systems the environment are all detectable through the robot’s chooses high strength rules, or rules that are in the sensors. vicinity of high strength ones. This means that * Crisp conditions are initially sufficient for learning. learning only takes place along the fringe of the state- * Disjunctions of conjunctions of conditions are space that has already been connected to a goal. While enough to characterise any state. animats should be allowed to seek knowledge that * Temporal Credit Allocation problem [5] may be may not have immediate use for the goal at hand. overlooked. The Learning Mechanism on the other hand is a * Actions are taken in serial. They are finite in system that learns incrementally (i.e. every rule is duration, and do not destabilise the system. built in a number of steps) using explicit units of * Relevant results can be pre-defined in terms of a representation (schemas). The algortihm aims to combination of conditions. enable robots to acquire and use sensorimotor * There are any number of agents in the environment. knowledge autonomously, without a priori knowledge Any one of them may be monitored by the learning of the skills concerned. This algorithm is based on the algorithm. Schema Mechanism developed by Drescher [4]. It was 2.2 Definiton of Terms altered and amended significantly in three main ways • Schema and Schema Space
  • 3. result follow more reliably. Reliability of a schema is context result measured by the ratio of: the number of times that its action extended structures action is executed, in the right context, and leads to the fulfilment of its result, to the total number of times schemai that its action is executed in the right context. Fig. 1 A schema. • Configuration Parameters and others The main structure of the learning mechanism is the - Result spin-off: a new schema made out of a copy of Schema (or rule) Space. The Schema Space is the a previous one, by adding a condition to the result collection of all schemas. At any time the Schema side. Space of the robot represents all its knowledge. The job of the learning algorithm is simply to create, - Context spin-off: a new schema made out of a copy modify, delete, and possibly link schemas. of a previous one, by adding a condition to the context side. A schema representation is made of two main structures: a Main Body (which comprises of a - θ1: the relevance threshold, for producing result context, action and result), and the extended spin-offs. structures. (see Fig. 1.) The main body is a tripartite - N1: the total number of experiments that need to be rule representing a counterfactual assertion. The taken before a result spin-off is allowed. extended structures keep information that is mainly used for creating (or spinning-off) new schemas. - θ2: the reliability threshold, used for producing context spin-offs. A schema has both declarative and procedural aspects. - N2: The number of activations that a result spin-off Declaratively, it is a unit of information about an schema needs to go through before it is allowed to interaction between the robot and the environment. produce a context spin-off. Procedurally, a schema represents a possible action to be taken at situations when its context is fulfilled and 3. The Main Algorithm its result is desired. The learning mechanism is best described by explaining the main algorithm that it embodies. This The components of a schema are: main algorithm goes through the following main • Main Body: steps: - Conditions: A condition may be viewed as a 1. Randomly select an action and execute it. function representing the degree of membership (or 2. Use the data collected before, during and after D.O.M.) of a sensor's output in a set representing that taking the action of the schema in its context, to condition. In a crisp DOM case, a condition can either update the two sets of correlation statistics be true or false. 3. Based on the statistics in step 2, the rule base may be updated as detailed in the algorithmic notation. - Context, Result and Action: A Context (and 4. Repeat steps 1 to 3 above until the predetermined similarly a Result) is a conjunction of one or more number of experiments is met. conditions (and their negations.) A Result could be either predefined or created at run-time. Contexts of The two phases of rule base update is best described in reliable schemas are automatically added, at run-time, the following algorithmic notation: to the set of results. An Action represents a command If ( no of experiments > N1 to an effector to take an action. If an Action is taken AND PTC/NTC(Resulti) >= θ1 then it's command is executed. AND Resulti not used before) • Extended Structures then Result spin-off Each schema has extended structures that contain two When update is completed and once the PSC and NSC main sets of correlation statistics. These statistics are are known context spin-off takes place according to: necessary for the development of schemas. The first If (no. of experiments > N2 set contains the Positive Transition Correlation (PTC) AND PSC/NSC(Conditioni) >= θ2 and the Negative Transition Correlation (NTC) are AND Conditioni not used before) used to find relevant results of an action. A relevant then Context spin-off result of an action is a result that has empirically shown that it follows the execution of that action 4. Problem and Experiments significantly more often than other actions from the The learning algorithm is now applied to the problem robot’s repertoire of actions. The PTC discovers of robot navigation. The goal of this application is to: positive results while the NTC discovers negative ones. The second set of statistics contains the Positive * Show that the algorithm is capable to deduce a list Success Correlation (PSC) and the Negative Success of rules that is capable (if properly pruned) of Correlation (NSC). PSC is used to find conditions that controlling the navigational behaviour of the robot when included in the context of a schema, will make navigating an obstacle-free path planned in a given its result follow more reliably than before adding environment. these conditions. NSC has the same function as PSC * Investigate the results of fuzzifying the except that it is used to find conditions that need to be context/result conditions on the execution of the excluded from the context of a schema to make its deduced rule base.
  • 4. 4.1 The Robot Simulation and the Task to Learn IF right_big ^ left_slight THEN right_small The robot has a cylindrical body, a differential drive IF right_small ^ left_slight THEN centre with two independently driven motorised wheels that IF centre ^ left_slight THEN left_small perform both the diving and steering. Four castors to support the mobile base on the flour (see Fig. 2.) IF left_small ^ left_slight THEN left_big C C They were found with different reliability values depending on the specific series of experiments taken. DW L DW R The rules produced by the learning algorithm are then pruned using the criteria of:- 1. Relevance to the goal (heading towards the goal), C C 2. High reliability. Fig. 2 A schematic of the Mobile Robot’s Base. The above rules become: Steering results from driving the left and right wheels at different speeds. This arrangement enables the robot IF right_big ^ left_slight THEN right_small to turn around its centre. IF right_small ^ left_slight THEN centre The robot is equipped with an on-board electronic With respect to rule block 1, the final list becomes compass, and odometry for localisation. (put in operational form): The robot requires two commands, linear speed and IF DirDif: right_big THEN DirOut: left_far change of direction. These are separated into individual rotational speed commands for the two IF right_big THEN left_slight driving motors, which are put in a velocity closed- IF right_small THEN left_far loop control. The global position control loop is IF right_small THEN left_slight closed by the feedback coming from the localisation system. IF centre THEN straight The task we want to learn is navigating our robot on a IF left_small THEN right_slight path consisting of straight line segments. The learnt IF left_small THEN right_far navigation rulebase should be able to control the robot to traverse the planned path smoothly. IF left_big THEN right_slight A simulation of the kinematics and dynamics of the IF left_big THEN right_far described mobile robot base was used for testing the For the second block of rules, those that are concerned learnt control rules. The robot simulation links to with the control of the linear velocity of the robot FuzzyTECH 3.1 development environment [6], where when heading towards a goal, a number of constraints the rules and the input/output membership functions is placed on the learning algorithm:- (including crisp ones) can be graphically edited. 1. Due to inertia, the robot is prevented from taking an 4.2 Experimental set-up for learning experiment in which the speed changes suddenly from The learning algorithm goes through two runs. One to slow to high or high to zero. Speed can only change discover the block of rules that are relevant to the gradually. This corresponds to real robots with orientation control. The second block contains rules dynamics, as opposed to mere kinematic simulations. that control the linear velocity of the robot. 2. We prune the first list of rules according to a different criteria (from the case with the orientation The learning algorithm is configured as follows: θ1:= block). This criteria is: 2, N1:= the total number of experiments taken, θ2:= 1, N2:= 3. Negative spin-off mechanisms are disabled. A. Highest reliability. B. Maximum distance traversal at each step. The sets of conditions and actions used are: C. Zero speed at the goal. DirDif={right_big, right_small, centre, left_small, This gives us the following list of rules: left_big}, Dist={very_near, near, medium, far}, SpIn={zero, slow, medium, high}, IF Dist: X ^ SpIn: zero THEN SpOut: slow DirOut={left_far, left_slight, straight, right_slight, IF medium ^ medium THEN medium right_far}, SpOut={ zero, slow, medium, high }. IF far ^ medium THEN high 5. Results IF far ^ slow THEN medium 5.1 Learning Algorithm Results IF far ^ high THEN high A series of experiments are fed to the learning algorithm. These experiments were chosen such that IF medium ^ high THEN medium they cover, on a uniformly random basis, the context IF medium ^ slow THEN medium space of the actions concerned. The learning IF near ^ slow THEN slow algorithm is run and a series of rules are produced. If the direction control action left_slight is taken as an IF very_near ^ slow THEN zero example we will find the following rules are produced:
  • 5. Since the two blocks of rules are learned separately, However, when appropriate fuzzy membership separation is enforced in action. This is done via functions replace the crisp ones, the performance of adding another block of rules which makes sure that the learnt navigation rules significantly improves, as the speed rules are only active when the robot is Fig. 4 shows. This is because when the robot becomes heading towards the goal. This special block is: closer to the direction of the goal, the final output of the orientation control rules is significantly reduced IF DirOut: right_far THEN SpOut: zero according to the degree of fulfilment. IF left_far THEN zero 6. Conclusions and Recommendations IF right_slight THEN zero The learning algorithm succeeded in finding the IF left_slight THEN zero declarative rules that represent, in their totality, the This means that at execution the robot should first interaction between the robot and the environment. execute the first set making sure that robot is in the Many of these rules were operationally useless, and right direction and then the second block starts had to be pruned (according to the criteria mentioned executing. previously). Once pruned, the resulting rules (both in crisp and fuzzy forms) were effective in controlling 5.2 Simulation Results the robot in navigation. We have shown that the performance of the learnt schemas improves as the context conditions are fuzzified. Hence, our future work would be making the learning schema mechanism a fuzzy one, which would be more general and capable of learning tasks in the continuous real world. Our learning mechanism, presented in this paper, readily allows this extension. 7. References [1] Dorigo M. Et al. (1994) "A comparison of Q-learning and classifier systems." In From animals to animats 3, edited by D.Cliff et al. MIT Press, Cambridge, MA. [2] Rescorla R. (1990) "Evidence for an association between the disciminative stimulus and the response-outcome association in instrumental learning." Journal of experimental psychology: animal behavior process, 16, 326-334. Fig. 3 Navigation using Crisp Conditions. [3] Roitblat H.(1994) "Mechanism and process in animal behavior: models of animals, animals as models." In From animals to animats Fig. 3 shows the robot navigating a planned path using 3, edited by D.Cliff et al. MIT Press, Cambridge, MA. the learnt rules with crisp membership functions for [4] Drescher G. (1990) "Made-up minds: a constructivist approach to the context conditions. It is clear that the robot’s artificial intelligence." MIT Press, Cambridge, MA. centre moved off the straight path segments, due to [5] Holland J H. (1992) “Adaptation in Natural and Artificial non-overlapping between the straight and the Systems.” MIT Press, Cambridge, MA. contiguous membership functions, and to its width. [6] FuzzyTECH 3.1 Software Manuals. This is unsuitable in cluttered environments (e.g. a narrow corridor). Had the straight membership function been narrower the robot would have swung right and left of the path in a zigzag, due to the activation of exactly the same rules regardless of the required amount of direction change when change of direction is required. Fig. 4 Navigation using Fuzzy Conditions.