Environments & Agents

Introduction to AI (I2AI)

Andy Weeger

Neu-Ulm University of Applied Sciences

March 12, 2025

Agents

Agency

Agency is the capacity of a system to maintain a continuous feedback loop with its environment. Agency requires a mapping of a history of environmental percepts to a sequence of actions designed to achieve a goal or maximize a performance measure (Russel & Norvig, 2022).

Core components

To analyze these systems, we define the following core components:

  • Agent: anything that perceives its environment through sensors and acts upon it through actuators.
  • Percept: The agent’s perceptual inputs at any given instant.
  • Percept Sequence: The complete history of everything the agent has ever perceived; an agent’s choice of action depends on this sequence observed to date.
  • Sensors: Mechanisms (cameras, GPS, microphones) that receive environmental input.
  • Actuators: Mechanisms (wheels, display screens, robotic joints) that execute actions.

Architecture

We must distinguish between the abstract logic and the physical implementation:

  • Agent function: An abstract mathematical mapping (\(f: \mathcal{P}^* \to \mathcal{A}\)) that describes how any given percept sequence results in an action.
  • Agent program: The concrete physical implementation (the actual code) running on a specific architecture.

Rational agents

Figure 1: Rational agents interact with environments through sensors and actuators

Exercise

Under which circumstances does a vacuum cleaning agent act rational?

Performance measure

If we use, to achieve our purposes, a mechanical agency with those operation we cannot interfere once we have started it […] we had better be quite sure that the purpose built into the machine is the purpose which we really desire Wiener (1960, p. 1358)

It is difficult to formulate a performance measure correctly.
This is a reason to be careful.

Rationality vs. perfection

Rationality is not the same as perfection.

  • Rationality maximizes expected performance.
  • Perfection maximizes actual performance.
  • Perfection requires omniscience.
  • Rational choice depends only on the percept sequence to date.

Comparison of performance standards

To understand the engineering limits of AI, we distinguish between three standards:

Metric Definition Info Requirement Feasibility
Rationality Maximizing expected performance Percept sequence + prior knowledge High: The engineering standard
Omniscience Knowing the actual outcome of actions Complete future and present data Impossible: Requires a “crystal ball”
Perfection Maximizing actual performance Requires Omniscience Impossible in unpredictable worlds

Overcoming ignorance

To bridge the gap between initial ignorance and rational behavior, agents must utilize information gathering and learning.

Since agents lack omniscience, they must be designed to:

  • Information gathering: take actions specifically to modify future percepts (e.g., looking both ways before crossing a street).
  • Learning: modify their internal agent function based on experience to improve performance over time.

As the environment is usually not completely known a priori and not completely predictable, these are vital parts of rationality (Russel & Norvig, 2022, p. 59).

Environments

Components

Before designing an agent (i.e., the solution), the task environment (i.e., the problem) must be specified as fully as possible, including

  • the performance measure,
  • the environment,
  • the actuators, and
  • the sensors

Russel & Norvig (2022) uses the short form PEAS to describe these parts of the task environment.

Properties

Task environments can be categorized along following dimensions:

  • Fully observable ↔︎ partially observable
  • Single agent ↔︎ multi-agent
  • Deterministic ↔︎ nondeterministic
  • Episodic ↔︎ sequential
  • Static ↔︎ dynamic
  • Discrete ↔︎ continuous
  • Known ↔︎ unknown

The hardest case is partially observable, multi-agent, nondeterministic, sequential, dynamic, and continuous (Russel & Norvig, 2022, pp. 62–64).

Exercise

Describe the task environment of a taxi driver agent.

Agent types

Simple reflex agents

Figure 2: A simple reflex agent1

Model-based reflex agents

Figure 3: A model-based reflex agent

Goal-based agents

Figure 4: A model-based, goal-based agent

Utility-based agents

Figure 5: A model-based, utility-based agent

Recap

What are the main differences between the agents?

Learning agents

Figure 6: A learning agent

On rationality

A rational agent is one
that does the right thing.

Utility-based learning agents are rational agents as they act so as to achieve the best outcome or, when there is uncertainty, the best expected outcome. This means that for each possible percept sequence, a rational agent should select an action that is expected to maximize its performance measure, given the evidence provided by the percept sequence and whatever built-in knowledge the agent has, which evolves over time (Russel & Norvig, 2022, p. 58).

Evolution of agents

The evolution of AI agents

 

 

 

Agentic AI

Definition

Agentic AI is an emerging paradigm in AI that refers to autonomous systems designed to pursue complex goals with minimal human intervention. Acharya et al. (2025, p. 18912)

Core characteristics of Agentic AI are

  • higher autonomy and goal complexity,
  • ability to adapt to environmental and situational unpredictabilities, and
  • independent decision-making.

Comparison with traditional AI

Comparison of traditional AI and Agentic AI based on Acharya et al. (2025)
Feature Traditional AI Agentic AI
Primary purpose Task-specific automation Goal-oriented autonomy
Human intervention High (predefined parameters) Low (autonomous adaptability)
Adaptability Limited High
Environment interaction Static or limited context Dynamic and context-aware
Learning type Primarily supervised Reinforcement and self-supervised
Decision-making Data-driven, static rules Autonomous, contextual reasoning

Comparison of agent types

Comparison between classical agents, reinforcement learning agents, and agentic AI based on Acharya et al. (2025)
Feature Classical Agents Learning Agents Agentic AI
Primary Purpose Fixed-task automation Reward-driven optimization Goal-oriented autonomy
Adaptability Low Moderate High
Learning Type Supervised Reinforcement Learning Hybrid, including RAG and Memory
Applications Static systems Dynamic environments Complex, multi-objective tasks

Q&A

Exercises

Concepts

Define in your own words the following terms:

  • Rationality
  • Autonomy
  • Agent
  • Environment
  • Sensor
  • Actuator
  • Percept
  • Agent function
  • Agent program

Agent types

Explain the differences between the following agent types in your own words. Describe the component(s) that is/are specific for each type.

  • Reflex agent
  • Model-based agent
  • Goal-based agent
  • Utility-based agent
  • Learning agent

Vacuum cleaner

Under which circumstances does a robotic vacuum cleaner act rational?

Describe the task environment of such an agent.

PEAS

For each of the following agents, specify the performance measure, the environment, the actuators, and the sensors.

  • Microwave oven
  • Chess program
  • Autonomous supply delivery

Performance measure

Describe a task environment in which the performance measure is easy to specify completely and correctly, and a in which it is not.

Assertions

For each of the following assertions, say whether it is true or false and support your answer with examples or counterexamples where appropriate.

  1. An agent that senses only partial information about the state cannot be perfectly rational.
  2. There exist task environments in which no pure reflex agent can behave rationally.
  3. There exists a task environment in which every agent is rational.
  4. Every agent is rational in an unobservable environment.
  5. A perfectly rational poker-playing agent never loses.

Task environment

For each of the following activities characterize the task environment it in terms of the properties discussed in the lecture notes.

  • Playing soccer
  • Exploring the subsurface oceans of Titan
  • Shopping for used AI books on the internet
  • Playing a tennis match

Task environment #2

For each of the following task environment properties, rank the example task environments from most to least according to how well the environment satisfies the property.

Lay out any assumptions you make to reach your conclusions.

  1. Fully observable: driving; document classification; tutoring a student in calculus; skin cancer diagnosis from images
  2. Continuous: driving; spoken conversation; written conversation; climate engineering by stratospheric aerosol injection
  3. Stochastic: driving; sudoku; poker; soccer
  4. Static: chat room; checkers; tax planning; tennis

Literature

Acharya, D. B., Kuppan, K., & Divya, B. (2025). Agentic AI: Autonomous intelligence for complex goals–a comprehensive survey. IEEE Access.
Anthrophic. (2024). Building effective agents. Anthropic Research Team; https://www.anthropic.com/engineering/building-effective-agents.
Russel, S., & Norvig, P. (2022). Artificial intelligence: A modern approach. Pearson Education.
Wiener, N. (1960). Some moral and technical consequences of automation. Science, 131(3410), 1355–1358.

Footnotes

  1. Rectangles are used to denote the current internal state of the agent’s decision process, rectangles with rounded corners to represent the background information used in the process.