Agents Flashcards

1
Q

Agents

A

Anything that can be viewed as perceiving its environment through sensors and acting upon that environment through actuators

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Percepts

A

Inputs obtained through any sensor (e.g. camera) which the agent perceives the environment through

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Actuators

A

Tools that the agent uses to act upon the environment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Abstract description of an agent, external

A

The agent function maps from percept histories to actions: [f: P* -> A]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Concrete implementation of an agent, internal

A

The agent program runs on physical architecture to produce f : agent = architecture + program

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Rational Agent

A

For each possible percept sequence, a rational agent should select an action that is expected to maximize its performance measure, given the evidence provided by the percept sequence and whatever built-in knowledge the agent has.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Rationality at any moment depends on:

A
  • the performance measure that defines the criterion of success
  • the agents prior knowledge of the environment
  • the actions that the agent can perform
  • the agents percept sequence to date
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Exploration as it relates to Rationality

A

to do with actions in order to modify the future percepts so that the agent would not do any uninformed action

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Learning as it relates to Rationality

A

Adapting to changing environment.

A successful agent computes its agent function in 3 periods:

1) when it is designed by its designer
2) when it decides its next action
3) when it learns from experiences to decide how to modify its behavior (improve the rules)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Autonomy as it relates to Rationality

A

it should learn what it can to compensate for partial or incorrect prior knowledge, relies more on its percepts than on the prior knowledge of its designer

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Task Environment

A

The problem to which the rational agent is a solution. Environment type largely determines agent design.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

PEAS

A

Performance Measure
Environment
Actuators
Sensors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Environment types:

Fully Observable

A

(vs. partially observable)

An agent’s sensors give it access to the complete state of the environment at each point in time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Environment types:

Deterministic

A

(vs. stochastic)

The next state of the environment is completely determined by the current state and the action executed by the agent

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Environment types:

Strategic

A

The environment is deterministic except for the actions of other agents

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Environment types:

Episodic

A

(vs. sequential)
The agents experience is divided into atomic “episodes” (each episode consists of the agent perceiving and then performing a single action), and the choice of action in each episode depends only on the episode itself

17
Q

Environment types:

Static

A

(vs. dynamic)

The environment is unchanged while an agent is deliberating

18
Q

Environment types:

Semi-Dynamic

A

The environment itself does not change with the passage of time, but the agent’s performance score does

19
Q

Environment types:

Discrete

A

(vs. Continuous)

A limited number of distinct, clearly defined percepts and actions

20
Q

Environment Types:

Single agent

A

(vs. multi-agent)

An agent operating by itself in an environment

21
Q

An agent is completely specified by…

A

… the agent function mapping percept sequences to actions.

22
Q

Percept sequence summarizes….

A

…the history of what the agent has perceived.

23
Q

An agent program is …

A

… an implementation or execution of the agent function.

  • The agent function is realized by an agent program executing on the physical architecture.
  • We aim to find an optimal agent program for the rational agent function.
24
Q

The four basic agent types, in order of increasing generality:

A

Simple Reflex Agents

Model-based Reflex Agents

Goal-based Reflex Agents

Utility-Based Agents

These can all be turned into Learning Agents

25
Q

Simple Reflex Agents

A
  • Selects actions on the basis of the current percept, ignoring the rest of the percept history
  • Some input has an established connection in the agent program to an action: such a connection is called a “condition- action rule” AKA if-then rules (if car in front breaking, then initiate breaks)
  • Simple, but limited intelligence
  • Will work only if correct decision can be made on the basis of just the current percept (only if the environment is fully observable)
  • Infinite Loops often unavoidable
26
Q

Model-based Reflex Agents

A
  • Agent maintains some internal state that depends on percept history, reflects at least some of the unobserved aspects of the current state.
  • Requires ‘Transition Model’: effects of agents actions, and how the world evolves independently from the agent
  • Requires ‘Sensor Model’: information about how the state of the world is reflected in the agents percepts
  • ## Seldom can determine current state of a partially observable environment exactly - represents agent’s “best guess”. Uncertainty unavoidable.
27
Q

Goal-based Agent

A
  • Keeps track of world state, as well as set of goals it is trying to achieve, and chooses an action that will (eventually) lead to the achievement of its goals
  • Search and Planning AI subfields deal with finding action sequences that achieve an agent’s goal
  • Flexible, because the knowledge that supports its decisions is represented explicitly and can be modified.
28
Q

Utility-Based Agent

A
  • Uses a model of the world, along with a utility function that measures its preferences among states of the world.
  • Then it chooses the action that leads to the best expected utility, where expected utility is computed by averaging over the all possible outcomes, weighted by the probability of the outcome.
  • Utility refers to the quality of being useful
  • An agent’s utility function is essentially an internalization of a performance measure
  • Complicated to design - has to keep track of environment, and choosing the utility-maximizing function is a difficult task
29
Q

Learning Agents

A
  • A process of modification of each component of the agent to bring the components into closer agreement with the available feedback information, thereby improving the overall performance of the agent.
  • Any type of agent (model-based, goal-based, utility-based) can be built as a learning agent (or not).
  • Advantage, allows agent to operate in initially unknown environments and to become more competent than its initial knowledge alone might allow

4 components:

  • Performance Element
  • Learning Element
  • Critic
  • Problem Generator
  • (outside component) Performance Standard
30
Q

Performance Element

A

Responsible for selecting external actions. What is considered the entire agent in other models: it takes in percepts and decides on actions.

31
Q

Learning Element

A

Responsible for making improvements, takes feedback from ‘critic’ on how the agent is doing and determines how the performance element could be modified to do better.

Design of Learning Element depends on design of performance element.

32
Q

Critic

A
  • Tells the Learning element how well the agent is doing with respect to a fixed performance standard.
  • Precepts themselves provide no indication of the agent’s success.
33
Q

Problem Generator

A
  • Responsible for suggesting actions that will lead to new and informative experiences
  • Suggests exploratory actions to the agent, some exploration action that could do some perhaps suboptimal results in short term, but would help discover much better actions in the long run.
  • Might identify parts of the model that are in need of improvement and suggest experiments
34
Q

Performance Standard

A
  • Distinguishes part of the incoming percept as a reward or penalty that provides direct feedback on the quality of the agent’s behavior
  • It is important that the performance standard be fixed. Conceptually, think of it as being outside the agent altogether, because the agent must not modify it to fit its own behavior.
  • Precepts themselves provide no indication of the agent’s success.