310 likes | 833 Views
Intelligent Agent. Chapter 2. Outline. Agent and Environment Rationality Performance Measure Environment Type Agent Type. Agent interacting with Environment. Agents include Human, robots, softbots, thermostats, etc. The agent function maps from percept histories to actions
E N D
Intelligent Agent Chapter 2
Outline • Agent and Environment • Rationality • Performance Measure • Environment Type • Agent Type 2
Agent interacting with Environment • Agents include Human, robots, softbots, thermostats, etc. • The agent function maps from percept histories to actions • The agent program runs on the physical architecture to produce f 3
Fig 2.3 4
Rational Agent • Characteristics of Rational Agent • tries to maximize expected value of performance measure • performance measure = degree of success • on the basis of evidence obtained by percept sequence • using its built-in prior world knowledge • Rational Agent : agent that does the RIGHT things Rational not= Omniscient, clairvoyant, successful Right Decision vs. Lucky decision example : Playing lotto 5
Rationality • Rationality Information-Gathering, learning, autonomy • Information-Gathering • Modify future percepts • Exploration unknown environment • Learning • Modify prior knowledge • Autonomy • Learn to compensate for partial and incorrect prior knowledge • Become independent of prior knowledge • Successful in variety of environment importance of learning 6
Performance Measure, Environment, Actuators, Sensors • To design a rational agent, we must specify task environment which consists of PEAS (Performance Measure, Environment, Actuators, Sensors) • Taxi Driver • Performance measure : safety, fast, legal, confortable trip, maximize profits • Environment : Roads, other traffic, pedestrains, customers • Actuators : steering, accelerator, brake, signal, horn, display • Sensors : cameras, sonar, speedometer, GPS, odometer, accelerometer, engine sensors, keyboard or microphone to accept destination 7
Internet Shopping Agent • Performance Measure : • Environment : • Actuators : • Sensors : 8
Properties of Task Environments • Fully observable vs. Partially observable • Deterministic vs. Stochastic • Strategic : deterministic except for actions of other agents • Episodic vs. Sequential • Static vs. Dynamic • Discrete vs. Continuous • Single Agents vs. Multiagent • Competitive, cooperative 9
Task Environment Types Real world is … 10
2-3. Structure of Intelligent Agents perception action ? Agent = architecture + program 11
Types of agents • Four basic types • Simple Reflex Agent • Reflex Agent with state • that keeps track of the world • Also called model-based reflex agent • Goal-based agent • Utility-based agent • All these can be turned into Learning Agents Generality 12
(1) Simple Reflex Agent • characteristics • no plan, no goal • do not know what they want to achieve • do not know what they are doing • condition-action rule • If condition then action • architecture - [fig. 2.9] program - [fig. 2.10] 13
(2) Model-based Reflex Agents • Characteristics • Reflex agent with internal state • Sensor does not provide the complete state of the world. • Updating the internal world • requires two kinds of knowledge which is called model • How world evolves • How agent’s action affect the world • architecture - [fig 2.11] program - [fig 2.12] 16
(3) Goal-based agents • Characteristics • Action depends on the GOAL . (consideration of future) • Goal is desirable situation • Choose action sequence to achieve goal • Needs decision making • fundamentally different from the condition-action rule. • Search and Planning • Appears less efficient, but more flexible • Because knowledge can be provided explicitly and modified • Architecture - [fig 2.13] 18
(4) Utility-based agents • Utility function • Degree of happiness • Quality of usefulness • map the internal states to a real number • (e.g., game playing) • Characteristics • to generate high-quality behavior • Rational decisions are made • Looking for higher Utility value • Expected Utility Maximizer • Explore several goals • Structure - [fig 2.14] 20
Learning Agents • Improve performance based on the percepts • 4 components • Learning elements • Making improvement • Performance elements • Selecting external actions • Critic • Tells how well the agent doing based on fixed performance standard • Problem generator • Suggest exploratory actions 22