Artificial Intelligence: A Modern Approach Chapter 2 Intelligent Agents

Ace your homework & exams now with Quizwiz!

Utility Function

An internalization of an agent's performance measure

Agent Function

A mapping of a given percept sequence to an action.

Uncertain

A partially observable or stochastic environment.

Learning Agent

An agent that can compensate for partial or incorrect knowledge.

Goa-based agent

An agent that combines a world model with a goal that describes desirable states to make decisions.

Model-based reflex agents

An agent that keeps track of unobservable aspects of the current state using the percpet to develop a model.

Percept

An agent's perceptual inputs at any given instant

Non-deterministic

An environment in which actions are characterized by their possible outcomes, but no probabilities are attached to them.

Agent

Anything that can be view as perceiving its environment through sensors and acting upon that environment through actuators.

Learning

As the agent gains experience, its initial configuration could be modified or augmented

Expressiveness

Complexity of learning and reasoning increases with expressiveness. The Conciseness of states increases with expressiveness. More expressive languages can capture at least as much information as less expressive ones but more concisely and complexly.

Actuators

Control of actions, production of results

Information gathering

Doing actions in order to modify future percepts.

Atomic representation

Each state of the world is indivisible, it has no internal structure.

Episodic/Sequential

Experience can be divided into atomic episodes. In each episode the agent receives a percept and performs an action. The next episode does not depend on the previous one.

Discrete/Continuous

If an agent's action and percepts and the state can have an infinite number of values at any given time then the environment is continuous.

Observable

If an agent's sensors give it access to the complete state of the environment as each point in time, then we say that the task environment is fully observable. (Fully/Partially)

Static/Dynamic

If the environment can change while the agent is deliberating then it is dynamic.

Stochastic/Deterministic

If the next state of the environment is completely determined by the current state and the action executed by the agent, then is is deterministic.

Known/unknown

In a known environment, the outcomes for all actions are given.

Problem Generator

It is responsible for suggesting actions that will lead to new and informative experiences.

Single/Multi-agent

Number of agents in an environment as well as the manner in which they interact (competitive, cooperative).

Sensors

Receiving input from all sources necessary to solve the problem.

Critic

Responsible for determining the how the agent is doing and modifying the performance element.

Learning Element

Responsible for making improvements

Performance Element

Responsible for selecting external actions, it takes in percpets and decides on actions.

Simple reflex agent

Selects actions on the basis of the current percpt, ignoring the rest of the percept history.

Factored representation

Splits up each state into a fixed set of variable or attributes, each of which can have a value.

Autonomy

The ability to learn how to compensate for partial or incorrect prior knowledge

Percept Sequence

The complete history of everything the agent has ever perceived.

Semidynamic

The environment does not change but an agent's performance does.

Performance Measure

The evaluation of the desirability of any given sequence of environment states.

Expected Utility

The expected value of an action's outcome in a partially unobservable or stochastic environment.

Environment

The external forces or actors acting on, changing, or causing a problem.

Agent Program

The internal implementation of an agent's agent function.

Utility

The measure of desirability

Rationality

The performance measure that defines the criterion of success. The agent's prior knowledge of the environment. The actions that the agent can perform. The agent's percept sequence to date. For each possible percept sequence, a rational agent should select an action that is expected to maximize its performance measure, given the evidence provided by the percept sequence and whatever built-in knowledge the agent has.

Task environments

The problems to which rational agents are the solution. PEAS (Performance Measures, Environment, Actuators, Sensors).

Structured representation

each state consists of objects which may have attributes of the their own and relationships to other objects


Related study sets

Chapter 2: Client care and Body Systems - Practice Exercise

View Set

Chapter 24: Using Nursing Research in Practice: 8th edition

View Set

BACTERIAL CAUSES OF SORE THROAT I: Streptococcus

View Set

ACT 101 Chapter 9 Cash Receipts, Cash Payments, and Banking Procedures

View Set

ATP 7-22.01: HOLISTIC HEALTH AND FITNESS TESTING

View Set

A03 - eLearning - Math: Fractions and Decimals 111

View Set