SlideShare a Scribd company logo
1 of 12
COMMAND AGENTS THAT MAKE HUMAN-LIKE DECISIONS FOR NEW TACTICAL SITUATIONS Masood Raza and Dr  Venkat V S S Sastry
RPD architecture
Soar Cognitive architecture
RPD and Soar Cognitive architecture
RPD-Soar Agent four inputs Three hidden layers, 12 neurons # known  situations
Speaker notes ,[object Object],[object Object],[object Object]
Experiment – Environment Use NN for recognizing the situation,  not to  produce a plan directly
Experiment - 1 ,[object Object]
Experiment - 2 ,[object Object],[object Object],[object Object],[object Object]
Analysis  ,[object Object]
Analysis ,[object Object],[object Object]
Summary ,[object Object],[object Object],[object Object],[object Object]

More Related Content

Viewers also liked

Semantical Cognitive Scheduling
Semantical Cognitive SchedulingSemantical Cognitive Scheduling
Semantical Cognitive Schedulingigalshilman
 
Kato Mivule: COGNITIVE 2013 - An Overview of Data Privacy in Multi-Agent Lear...
Kato Mivule: COGNITIVE 2013 - An Overview of Data Privacy in Multi-Agent Lear...Kato Mivule: COGNITIVE 2013 - An Overview of Data Privacy in Multi-Agent Lear...
Kato Mivule: COGNITIVE 2013 - An Overview of Data Privacy in Multi-Agent Lear...Kato Mivule
 
Multi agent paradigm for cognitive parameter based feature similarity for soc...
Multi agent paradigm for cognitive parameter based feature similarity for soc...Multi agent paradigm for cognitive parameter based feature similarity for soc...
Multi agent paradigm for cognitive parameter based feature similarity for soc...eSAT Journals
 
Stufano, Borri and Rabino - input2012
Stufano, Borri and Rabino - input2012Stufano, Borri and Rabino - input2012
Stufano, Borri and Rabino - input2012INPUT 2012
 
Enhancing Intelligent Agents By Improving Human Behavior Imitation Using Sta...
Enhancing Intelligent Agents By Improving Human Behavior  Imitation Using Sta...Enhancing Intelligent Agents By Improving Human Behavior  Imitation Using Sta...
Enhancing Intelligent Agents By Improving Human Behavior Imitation Using Sta...Osama Salaheldin
 
Depth of Feelings: Modeling Emotions in User Models and Agent Architectures
Depth of Feelings: Modeling Emotions in User Models and Agent ArchitecturesDepth of Feelings: Modeling Emotions in User Models and Agent Architectures
Depth of Feelings: Modeling Emotions in User Models and Agent ArchitecturesEva Hudlicka
 
The Design of Cognitive Social Simulation Framework using Statistical Methodo...
The Design of Cognitive Social Simulation Framework using Statistical Methodo...The Design of Cognitive Social Simulation Framework using Statistical Methodo...
The Design of Cognitive Social Simulation Framework using Statistical Methodo...IJORCS
 
Artwork creation by a cognitive architecture integrating computational creati...
Artwork creation by a cognitive architecture integrating computational creati...Artwork creation by a cognitive architecture integrating computational creati...
Artwork creation by a cognitive architecture integrating computational creati...Agnese Augello
 
Cognitive assistance at work
Cognitive assistance at workCognitive assistance at work
Cognitive assistance at workHamid Motahari
 
A Cloud-Based Bayesian Smart Agent Architecture for Internet-of-Things Applic...
A Cloud-Based Bayesian Smart Agent Architecture for Internet-of-Things Applic...A Cloud-Based Bayesian Smart Agent Architecture for Internet-of-Things Applic...
A Cloud-Based Bayesian Smart Agent Architecture for Internet-of-Things Applic...Veselin Pizurica
 
Extending the Mind with Cognitive Prosthetics?
Extending the Mind with Cognitive Prosthetics? Extending the Mind with Cognitive Prosthetics?
Extending the Mind with Cognitive Prosthetics? PhiloWeb
 
Cognitive Computing and the future of Artificial Intelligence
Cognitive Computing and the future of Artificial IntelligenceCognitive Computing and the future of Artificial Intelligence
Cognitive Computing and the future of Artificial IntelligenceVarun Singh
 
Cognitive behavior therapy theory and practice
Cognitive behavior therapy theory and practiceCognitive behavior therapy theory and practice
Cognitive behavior therapy theory and practiceWuzna Haroon
 
Cognitive learning theory
Cognitive learning theoryCognitive learning theory
Cognitive learning theoryVerbena Wilburg
 
A Novel Model of Cognitive Presence Assessment Using Automated Learning Analy...
A Novel Model of Cognitive Presence Assessment Using Automated Learning Analy...A Novel Model of Cognitive Presence Assessment Using Automated Learning Analy...
A Novel Model of Cognitive Presence Assessment Using Automated Learning Analy...Vitomir Kovanovic
 
Piaget's Cognitive Development
Piaget's Cognitive DevelopmentPiaget's Cognitive Development
Piaget's Cognitive DevelopmentARUL LAWRENCE
 
Routing algorithm
Routing algorithmRouting algorithm
Routing algorithmBushra M
 

Viewers also liked (19)

Semantical Cognitive Scheduling
Semantical Cognitive SchedulingSemantical Cognitive Scheduling
Semantical Cognitive Scheduling
 
Kato Mivule: COGNITIVE 2013 - An Overview of Data Privacy in Multi-Agent Lear...
Kato Mivule: COGNITIVE 2013 - An Overview of Data Privacy in Multi-Agent Lear...Kato Mivule: COGNITIVE 2013 - An Overview of Data Privacy in Multi-Agent Lear...
Kato Mivule: COGNITIVE 2013 - An Overview of Data Privacy in Multi-Agent Lear...
 
Multi agent paradigm for cognitive parameter based feature similarity for soc...
Multi agent paradigm for cognitive parameter based feature similarity for soc...Multi agent paradigm for cognitive parameter based feature similarity for soc...
Multi agent paradigm for cognitive parameter based feature similarity for soc...
 
Stufano, Borri and Rabino - input2012
Stufano, Borri and Rabino - input2012Stufano, Borri and Rabino - input2012
Stufano, Borri and Rabino - input2012
 
Enhancing Intelligent Agents By Improving Human Behavior Imitation Using Sta...
Enhancing Intelligent Agents By Improving Human Behavior  Imitation Using Sta...Enhancing Intelligent Agents By Improving Human Behavior  Imitation Using Sta...
Enhancing Intelligent Agents By Improving Human Behavior Imitation Using Sta...
 
Depth of Feelings: Modeling Emotions in User Models and Agent Architectures
Depth of Feelings: Modeling Emotions in User Models and Agent ArchitecturesDepth of Feelings: Modeling Emotions in User Models and Agent Architectures
Depth of Feelings: Modeling Emotions in User Models and Agent Architectures
 
The Design of Cognitive Social Simulation Framework using Statistical Methodo...
The Design of Cognitive Social Simulation Framework using Statistical Methodo...The Design of Cognitive Social Simulation Framework using Statistical Methodo...
The Design of Cognitive Social Simulation Framework using Statistical Methodo...
 
Artwork creation by a cognitive architecture integrating computational creati...
Artwork creation by a cognitive architecture integrating computational creati...Artwork creation by a cognitive architecture integrating computational creati...
Artwork creation by a cognitive architecture integrating computational creati...
 
Cognitive assistance at work
Cognitive assistance at workCognitive assistance at work
Cognitive assistance at work
 
A Cloud-Based Bayesian Smart Agent Architecture for Internet-of-Things Applic...
A Cloud-Based Bayesian Smart Agent Architecture for Internet-of-Things Applic...A Cloud-Based Bayesian Smart Agent Architecture for Internet-of-Things Applic...
A Cloud-Based Bayesian Smart Agent Architecture for Internet-of-Things Applic...
 
Extending the Mind with Cognitive Prosthetics?
Extending the Mind with Cognitive Prosthetics? Extending the Mind with Cognitive Prosthetics?
Extending the Mind with Cognitive Prosthetics?
 
Cognition consumer behaviour
Cognition consumer behaviourCognition consumer behaviour
Cognition consumer behaviour
 
Cognitive Computing and the future of Artificial Intelligence
Cognitive Computing and the future of Artificial IntelligenceCognitive Computing and the future of Artificial Intelligence
Cognitive Computing and the future of Artificial Intelligence
 
Cognitive behavior therapy theory and practice
Cognitive behavior therapy theory and practiceCognitive behavior therapy theory and practice
Cognitive behavior therapy theory and practice
 
Cognitive learning theory
Cognitive learning theoryCognitive learning theory
Cognitive learning theory
 
A Novel Model of Cognitive Presence Assessment Using Automated Learning Analy...
A Novel Model of Cognitive Presence Assessment Using Automated Learning Analy...A Novel Model of Cognitive Presence Assessment Using Automated Learning Analy...
A Novel Model of Cognitive Presence Assessment Using Automated Learning Analy...
 
Towards a "Mindful" Web
Towards a "Mindful" WebTowards a "Mindful" Web
Towards a "Mindful" Web
 
Piaget's Cognitive Development
Piaget's Cognitive DevelopmentPiaget's Cognitive Development
Piaget's Cognitive Development
 
Routing algorithm
Routing algorithmRouting algorithm
Routing algorithm
 

Similar to Command Agents That make Human Like Decisions F

DRL Medical Imaging Literature Review
DRL Medical Imaging Literature ReviewDRL Medical Imaging Literature Review
DRL Medical Imaging Literature ReviewJocelyn Baduria
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural networkNagarajan
 
coppin chapter 10e.ppt
coppin chapter 10e.pptcoppin chapter 10e.ppt
coppin chapter 10e.pptbutest
 
DagdelenSiriwardaneY..
DagdelenSiriwardaneY..DagdelenSiriwardaneY..
DagdelenSiriwardaneY..butest
 
Brains@Bay Meetup: The Effect of Sensorimotor Learning on the Learned Represe...
Brains@Bay Meetup: The Effect of Sensorimotor Learning on the Learned Represe...Brains@Bay Meetup: The Effect of Sensorimotor Learning on the Learned Represe...
Brains@Bay Meetup: The Effect of Sensorimotor Learning on the Learned Represe...Numenta
 
[PR12] understanding deep learning requires rethinking generalization
[PR12] understanding deep learning requires rethinking generalization[PR12] understanding deep learning requires rethinking generalization
[PR12] understanding deep learning requires rethinking generalizationJaeJun Yoo
 
Using a Long Short-Term Memory Recurrent Neural Network (LSTM-RNN) to Classif...
Using a Long Short-Term Memory Recurrent Neural Network (LSTM-RNN) to Classif...Using a Long Short-Term Memory Recurrent Neural Network (LSTM-RNN) to Classif...
Using a Long Short-Term Memory Recurrent Neural Network (LSTM-RNN) to Classif...BASMAJUMAASALEHALMOH
 
Model-Based Reinforcement Learning @NIPS2017
Model-Based Reinforcement Learning @NIPS2017Model-Based Reinforcement Learning @NIPS2017
Model-Based Reinforcement Learning @NIPS2017mooopan
 
deepnet-lourentzou.ppt
deepnet-lourentzou.pptdeepnet-lourentzou.ppt
deepnet-lourentzou.pptyang947066
 
50134147-Knowledge-Representation-Using-Rules.ppt
50134147-Knowledge-Representation-Using-Rules.ppt50134147-Knowledge-Representation-Using-Rules.ppt
50134147-Knowledge-Representation-Using-Rules.pptssuserec53e73
 
Local search algorithms
Local search algorithmsLocal search algorithms
Local search algorithmsbambangsueb
 
Artificial Neural Networks: Applications In Management
Artificial Neural Networks: Applications In ManagementArtificial Neural Networks: Applications In Management
Artificial Neural Networks: Applications In ManagementIOSR Journals
 
machinelearningengineeringslideshare-160909192132 (1).pdf
machinelearningengineeringslideshare-160909192132 (1).pdfmachinelearningengineeringslideshare-160909192132 (1).pdf
machinelearningengineeringslideshare-160909192132 (1).pdfShivareddyGangam
 

Similar to Command Agents That make Human Like Decisions F (20)

DRL Medical Imaging Literature Review
DRL Medical Imaging Literature ReviewDRL Medical Imaging Literature Review
DRL Medical Imaging Literature Review
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural network
 
coppin chapter 10e.ppt
coppin chapter 10e.pptcoppin chapter 10e.ppt
coppin chapter 10e.ppt
 
DagdelenSiriwardaneY..
DagdelenSiriwardaneY..DagdelenSiriwardaneY..
DagdelenSiriwardaneY..
 
Brains@Bay Meetup: The Effect of Sensorimotor Learning on the Learned Represe...
Brains@Bay Meetup: The Effect of Sensorimotor Learning on the Learned Represe...Brains@Bay Meetup: The Effect of Sensorimotor Learning on the Learned Represe...
Brains@Bay Meetup: The Effect of Sensorimotor Learning on the Learned Represe...
 
[PR12] understanding deep learning requires rethinking generalization
[PR12] understanding deep learning requires rethinking generalization[PR12] understanding deep learning requires rethinking generalization
[PR12] understanding deep learning requires rethinking generalization
 
Chapter 5 (final)
Chapter 5 (final)Chapter 5 (final)
Chapter 5 (final)
 
Nn devs
Nn devsNn devs
Nn devs
 
Using a Long Short-Term Memory Recurrent Neural Network (LSTM-RNN) to Classif...
Using a Long Short-Term Memory Recurrent Neural Network (LSTM-RNN) to Classif...Using a Long Short-Term Memory Recurrent Neural Network (LSTM-RNN) to Classif...
Using a Long Short-Term Memory Recurrent Neural Network (LSTM-RNN) to Classif...
 
Model-Based Reinforcement Learning @NIPS2017
Model-Based Reinforcement Learning @NIPS2017Model-Based Reinforcement Learning @NIPS2017
Model-Based Reinforcement Learning @NIPS2017
 
deepnet-lourentzou.ppt
deepnet-lourentzou.pptdeepnet-lourentzou.ppt
deepnet-lourentzou.ppt
 
50134147-Knowledge-Representation-Using-Rules.ppt
50134147-Knowledge-Representation-Using-Rules.ppt50134147-Knowledge-Representation-Using-Rules.ppt
50134147-Knowledge-Representation-Using-Rules.ppt
 
Neural networks
Neural networksNeural networks
Neural networks
 
Local search algorithms
Local search algorithmsLocal search algorithms
Local search algorithms
 
Artificial Neural Networks: Applications In Management
Artificial Neural Networks: Applications In ManagementArtificial Neural Networks: Applications In Management
Artificial Neural Networks: Applications In Management
 
AI: Learning in AI
AI: Learning in AI AI: Learning in AI
AI: Learning in AI
 
AI: Learning in AI
AI: Learning in AI AI: Learning in AI
AI: Learning in AI
 
OCT-20
OCT-20OCT-20
OCT-20
 
B42010712
B42010712B42010712
B42010712
 
machinelearningengineeringslideshare-160909192132 (1).pdf
machinelearningengineeringslideshare-160909192132 (1).pdfmachinelearningengineeringslideshare-160909192132 (1).pdf
machinelearningengineeringslideshare-160909192132 (1).pdf
 

More from eAssessment in Practice Symposium (13)

Venkat Sastry talking on MoLE Project A Global Technology Initiative
Venkat Sastry talking on MoLE ProjectA Global Technology InitiativeVenkat Sastry talking on MoLE ProjectA Global Technology Initiative
Venkat Sastry talking on MoLE Project A Global Technology Initiative
 
An Introduction to MATLAB with Worked Examples
An Introduction to MATLAB with Worked ExamplesAn Introduction to MATLAB with Worked Examples
An Introduction to MATLAB with Worked Examples
 
On an evaluation strategy for measurement of learning
On an evaluation strategy for measurement of learningOn an evaluation strategy for measurement of learning
On an evaluation strategy for measurement of learning
 
Caa2013 9 10-july2013
Caa2013 9 10-july2013Caa2013 9 10-july2013
Caa2013 9 10-july2013
 
Defence education past, present, future 2
Defence education past, present, future 2Defence education past, present, future 2
Defence education past, present, future 2
 
Defed 2012 Conf Lr 03
Defed 2012 Conf Lr 03Defed 2012 Conf Lr 03
Defed 2012 Conf Lr 03
 
E-Assessment in Practice, 13-14 June 2012, Shrivenham, UK
E-Assessment in Practice, 13-14 June 2012, Shrivenham, UKE-Assessment in Practice, 13-14 June 2012, Shrivenham, UK
E-Assessment in Practice, 13-14 June 2012, Shrivenham, UK
 
Cranfield Brochure And Programme 31 Oct2011
Cranfield Brochure And Programme 31 Oct2011Cranfield Brochure And Programme 31 Oct2011
Cranfield Brochure And Programme 31 Oct2011
 
Technology Enhanced Learning E Assessment 12 14 July 2011
Technology Enhanced Learning E Assessment 12 14 July 2011Technology Enhanced Learning E Assessment 12 14 July 2011
Technology Enhanced Learning E Assessment 12 14 July 2011
 
S Ict 16 Dec
S Ict 16 DecS Ict 16 Dec
S Ict 16 Dec
 
Grasp Hsc09
Grasp Hsc09Grasp Hsc09
Grasp Hsc09
 
X Part Hsc09 28 Oct 2009
X Part Hsc09 28 Oct 2009X Part Hsc09 28 Oct 2009
X Part Hsc09 28 Oct 2009
 
On e-Assessment
On e-AssessmentOn e-Assessment
On e-Assessment
 

Command Agents That make Human Like Decisions F

Editor's Notes

  1. motivation why human-like decisions? what are tactical situiations? The use of simulation in the military is increasing for training, systems analysis, systems acquisition, evaluation of doctrines and command decision making. As human-in-the-loop simulations are time and personnel intensive, therefore, the need to represent human behavior in military simulations is becoming imperative. Realistic human behavior representation can contribute to the development of more realistic and thus more useful military simulations. Human-like decision making based on correct situational awareness can add realism to human behavior representation for military simulations. Military decision making process ( MDMP ) is based on Multi-attribute utility analysis ( MAUA ), which is a classical decision making approach and forms the basis of decision making model of most present day military simulations. MDMP is very procedural and often cumbersome process. When under stress due to time pressure and in dynamic situations, military commanders use their experience and abbreviate this process and adopt a naturalistic decision making model described by Klein and associates as Recognition primed decision making ( RPD ) [2]. Recognition primed decision making model is suitable for experienced commanders [3]. The basic concept of RPD is the recognition of a situation in a changing context, recalling a suitable course of action for this situation, and then either implementing the course of action straight away or after mentally simulating it to find out if it will work properly [4].
  2. The simplest, and probably most common, case within the RPD model is Level 1 (Figure 1), where a decision-maker sizes up a situation, forms expectancies about what's going to happen next, determines the cues that are most relevant, recognizes the reasonable goals to pursue in the situation, recognizes a typical CoA that is likely to succeed and carries it out. Level 2 is a more difficult case, in which the decision-maker isn't certain about the nature of the situation. Perhaps some anomaly arises that violates expectancies and forces the decision-maker to question whether the situation is different from what it seems, or perhaps uncertainty might be present from the beginning. Here, decision-makers do deliberate about what's happening. Level 3 of RPD model is the case in which decision maker arrives at an understanding of a situation and recognizes a typical CoA and then evaluates it by mentally simulating what will happen when it is carried out. In this way, if he spots weaknesses in the plan, he can repair it and improve the plan, or throw it away and evaluate the next plausible action. The model has been tested in variety of applications including fireground command, battle planning, critical care nursing, corporate information management, and chess tournament play.
  3. Soar is a symbolic cognitive architecture for general intelligence. It has been used for creating intelligent forces for large and small scale military simulations. Soar is a forward chaining rule based system. Both the declarative and procedural knowledge are represented as production rules. The production rules are condition-action pair. The long term memory ( LTM ) is composed of production rules while the short term memory ( STM ) contains only declarative knowledge. STM in Soar is also the working memory ( WM ) that holds all the dynamic data structures in the form of an identifier, attribute and value triplet called working memory elements ( WME ). The value in a WME may also be an identifier connecting to another attribute, thereby, forming a connected graph. Unlike some rule based systems, Soar is a parallel rule matching and parallel rule firing system. Impasse in Soar is the architecturally detected lack of available knowledge. It has a single architecturally supported goal stack. Each sub-goal creates a sub-state to bring to bear the knowledge to achieve the goal. Soar uses this hierarchical sub-goaling for task decomposition. This capability may also be used to contextualize the situation.
  4. Soar provides a convenient framework to model most of the aspects of the RPD model. The elaboration phase in Soar decision cycle is used for situation awareness and the problem space based architecture, automatic sub-goaling and creation of sub-states due to impasse is used for mental simulation
  5. The environment is developed in the Java object oriented programming language, the RPD model is implemented in the Soar cognitive architecture and the agent and the environment are interfaced with Soar mark-up language ( SML ). A trained artificial neural network is also integrated with the agent architecture to enhance the ability of the agent in handling new situations. The experiences of the command agent are stored in the LTM in the form of production rules. The success values for the courses of action for specific situations are represented numerically. All atomic actions, such as move, turn, fire, etcetera, expected to be performed by the agent in a simulation are coded by the modeller. The selection of an action for a specific situation in pursuit of single or multiple goals based on corresponding success values is the task of the RPD-Soar agent which forms the behaviour of the agent. This behaviour emerges at the simulation run time.
  6. In rule based systems the antecedents of a production rule have to match exactly for the production to fire. If the current situation deviates from the conditions in the rule then the appropriate rule does not fire. Due to rule matching through an efficient algorithm like RETE and also advances in computer technology it is possible in Soar to add a large number of production rules to handle generalization. The RETE algorithm efficiently solves the many-to-many matching problem encountered when rules are matched to facts (Forgy, 1982). Writing large number of rules is possible but is not an efficient method of solving this problem. Alternate approaches like similarity-based generalization, fuzzy logic and artificial neural network can solve this problem in a more efficient way. In this implementation, an artificial neural network is used for situation recognition. There are two reasons for using an artificial neural network in this implementation: first, it has already been used for a similar task with promising results (Liang et al., 2001); second, it has the ability to automatically prioritize the situations according to their level of similarity.  The situations are fed to the trained artificial neural network which matches the new situation to one of the known situations and gives the agent a recognized situation. The recognized situation has the complete set or a subset of its four constituents that are goals, courses of action, cues, and expectations. The agent selects the course of action for the situation and implements it with the help of lower level actions selected through mental simulation if required. The neural network is trained for each agent based on the range of situations it is likely to face. Motivated from the work of Liang et al. (2001), the neural net is a multi-layered normal feed forward network. It consists of an input layer of four nodes, three hidden layers of twelve nodes each, and the number of nodes in the output layer depends upon the number of known situations. The number of nodes in the output layer varies from situation to situation. For the experiments conducted in this research the configuration of the input layer and the hidden layers is not changed but these layers may also be reconfigured if required.
  7. The vignette for this experiment is motivated from the work of Liang et al. (2001). The domain is a military ground based operation. A military commander of a troop of tanks consisting of three tanks selects a strategy to attack the enemy tank in the north. The terrain is simple and it can have 0, 1, or 2 passable hills. The terrain with two hills is shown in Figure 2; the enemy is represented with a red square in the north at location (0, 1) and own position is the blue circle in the south at the origin. The agent’s own starting position and the enemy position remains the same through out the experiment. The commander selects a strategy based on the decisions that whether to divide the troop of tanks in an assault group ( AG ) and a fire support ( FS ) group or to use them as one group only. The commander also selects the intermediate and the final locations or location of these groups or group which also dictate the route to be adopted by the group(s). Use NN for recognizing the situation rather than prediction of plan The neural net in this experiment is used for pattern recognition and not for plan generation because, as Liang et al. also realize, the option to generate plan directly from the trained neural network does not prove to be successful. In this experiment the target for NN training in each case is the numeric value of ‘1’ for the output node corresponding to the recognized situation and ‘0’ for the rest of the output nodes. It is assumed that such a clear difference between two target values will produce better results compared to the mixed target values for the output nodes corresponding to different strategies in the work of Liang et al. (2001). Moreover, there is potential advantage in this representation for an RPD model. The advantage in this design is that for a given situation the output node with the highest value is considered as the recognized situation and if the evaluation of the corresponding course of action through mental simulation is not promising then the output node with the second highest value may be considered. The basic situations and corresponding strategies in the work of Liang et al. (2001) is used but some strategies are modified and some more strategies are added in the training set for this experiment based on the knowledge of the author on the subject. The reason for the addition of examples in the training set is to improve the performance of the net in recognition of new situations which are related to number of training examples and also to sufficiently cover the problem space. Each case giving the locations of hills is associated to a situation and there is a corresponding strategy for each situation. The neural net only recognizes the situation and then the agent retrieves the corresponding strategy. The neural net part of the agent is required to recognize new situations produced in the environment. In order to test the ability of the trained neural net in recognizing new situations we developed a set of situations that exhaustively covers the problem space. In the battlefield, the locations of enemy and own positions are fixed. The situational variables are the locations of hills in the terrain. As the locations of the hills are changed in the terrain, new situations are generated. The new situations are the ones that are not included in the training examples and for which the agent has not been trained .
  8. The neural net in this experiment recognizes the situation and gives the plan to RPD-Soar agent which in turn produces a strategy. RPD-Soar agent implements the attack plan by selecting lower level actions, such as finding the way to the final locations of fire support and assault groups, to achieve the goal using mental simulation. In order to the explore the problem space we fixed one hill and moved the other hill on the given terrain. As the hill moves new situations are generated and the neural net generates suitable strategies for the new situations by recognizing known situations and the related strategies. One such experiemnt is shown in the next slide.  
  9. The fire support group in this Strategy is taking the advantage of the presence of hill. The agent considers it an acceptable plan when the blue force makes use of one out of two hills available in the area. This decision is based on the theory of satisficing and not optimizing. Satisficing is the hall mark of recognition primed decision making. Some times it happens so that if you discard an acceptable plan and try the strategy with the next highest recognition value then this new strategy is providing the force with even lesser advantage than the previous strategy. This situation under discussion clearly demonstrates this case. The agent in this case accepts the previous plan because it made use of one hill but if it did not accept it then the next Strategy in line is shown in the next slide. Human behaviour representation in military simulations is not sufficiently realistic, specially the decision making by synthetic military commanders. The decision making process lacks realistic representation of variability, flexibility, and adaptability exhibited by a single entity across various episodes. It is hypothesized that a widely accepted naturalistic decision model, suitable for military or other domains with high stakes, time stress, dynamic and uncertain environments, based on well established cognitive architecture can address some of these deficiencies. And therefore, we have developed a computer implementation of Recognition Primed Decision Making (RPD) model using Soar cognitive architecture and it is referred to as RPD-Soar agent in this paper. The results of the experiments on RPD-Soar agent clearly demonstrate the ability of the model to improve realism in representing human decision making behaviour by exhibiting the ability to recognise a situation in a changing context, handle new situations, flexible and adaptive  in the decision making process. The experiments account for variability within and across individuals. Due to the ability of the RPD-Soar agent to mentally simulate applicable courses of action it is possible for the agent to handle new situations by mentally simulating all available courses of action and selecting a suitable course of action out of them. In order to make quick decisions the number of courses of action to be mentally simulated need to be kept at a minimum. This is achieved by integrating a trained neural network with the RPD-Soar agent. This trained neural network prioritizes the available courses of action in the order of their suitability to the present situation based on its training examples. RPD-Soar agent evaluates the course of action and if it satisfices the agent implements it otherwise throws it away and selects the next course of action in the order. The proposed implementation is evaluated using prototypical scenarios arising in command decision making in tactical situations of military land operations.
  10. In most cases i.e., 10 out of 12 the neural net recognizes the situation correctly which means it identifies the closest known situation. But some times the hills in this new situation are so located that the assault and/or fire support group can not take advantage of the hills for cover from observation and fire from the enemy. This happens when the hills in the new situation are close to the known situation for which a strategy exists and is also recognized but the off set is such that the hill is in the south or east or west of the group's position and is not located in between the group position and the enemy. The RPD-Soar agent then evaluates the recognized situation and if it is not taking advantage of any of the hills present in the vicinity of the enemy (cover from enemy observation and fire is an overwhelming factor in selecting a strategy to implement) then it discards this strategy and tries the strategy which has the next highest recognition value. That was Strategy number 9 but the blue force is not taking any advantage of the hills therefore the agent discards the Strategy and tries the Strategy that has the next highest recognition value.
  11. Instead of straight away discarding the plan with the highest recognition value it should be modified and then evaluated again. And then if it still does not satisfice then it should be discarded and the next in line be evaluated.