Environments & Agents

Introduction to AI

Andy Weeger

Neu-Ulm University of Applied Sciences

March 12, 2025

Agents

Rational agents

Figure 1: Rational agents interact with environments through sensors and actuators

Exercise

Under which circumstances does a vacuum cleaning agent act rational?

Performance measure

If we use, to achieve our purposes, a mechanical agency with those operation we cannot interfere once we have started it […] we had better be quite sure that the purpose built into the machine is the purpose which we really desire Wiener (1960, 1358)

It is difficult to formulate a performance measure correctly.
This is a reason to be careful.

Rationality vs. perfection

Rationality is not the same as perfection.

  • Rationality maximizes expected performance.
  • Perfection maximizes actual performance.
  • Perfection requires omniscience.
  • Rational choice depends only on the percept sequence to date.

Environments

Components

Before designing an agent (i.e., the solution), the task environment (i.e., the problem) must be specified as fully as possible, including

  • the performance measure,
  • the environment,
  • the actuators, and
  • the sensors

Russel and Norvig (2022) uses the short form PEAS to describe these parts of the task environment.

Properties

Task environments can be categorized along following dimensions:

  • Fully observable ↔︎ partially observable
  • Single agent ↔︎ multi-agent
  • Deterministic ↔︎ nondeterministic
  • Episodic ↔︎ sequential
  • Static ↔︎ dynamic
  • Discrete ↔︎ continuous
  • Known ↔︎ unknown

The hardest case is partially observable, multi-agent, nondeterministic, sequential, dynamic, and continuous (Russel and Norvig 2022, 62–64).

Exercise

Describe the task environment of a taxi driver agent.

Agent types

Simple reflex agents

Figure 2: A simple reflex agent3

Model-based reflex agents

Figure 3: A model-based reflex agent

Goal-based agents

Figure 4: A model-based, goal-based agent

Utility-based agents

Figure 5: A model-based, utility-based agent

Recap

What are the main differences between the agents?

Learning agents

Figure 6: A learning agent

Rational agents

A rational agent is one
that does the right thing.

Utility-based learning agents are rational agents as they act so as to achieve the best outcome or, when there is uncertainty, the best expected outcome. This means that for each possible percept sequence, a rational agent should select an action that is expected to maximize its performance measure, given the evidence provided by the percept sequence and whatever built-in knowledge the agent has, which evolves over time (Russel and Norvig 2022, 58).

Evolution of agents

The evolution of AI agents

 

 

 

Agentic AI

Definition

Agentic AI is an emerging paradigm in AI that refers to autonomous systems designed to pursue complex goals with minimal human intervention. Acharya, Kuppan, and Divya (2025, 18912)

Core characteristics of Agentic AI are

  • Higher autonomy and goal complexity
  • Ability to adapt to environmental and situational unpredictabilities
  • Independent decision-making

Comparison with traditional AI

Comparison of traditional AI and Agentic AI based on Acharya, Kuppan, and Divya (2025)
Feature Traditional AI Agentic AI
Primary purpose Task-specific automation Goal-oriented autonomy
Human intervention High (predefined parameters) Low (autonomous adaptability)
Adaptability Limited High
Environment interaction Static or limited context Dynamic and context-aware
Learning type Primarily supervised Reinforcement and self-supervised
Decision-making Data-driven, static rules Autonomous, contextual reasoning

Comparison of agent types

Comparison between classical agents, reinforcement learning agents, and agentic AI based on Acharya, Kuppan, and Divya (2025)
Feature Classical Agents Learning Agents Agentic AI
Primary Purpose Fixed-task automation Reward-driven optimization Goal-oriented autonomy
Adaptability Low Moderate High
Learning Type Supervised Reinforcement Learning Hybrid, including RAG and Memory
Applications Static systems Dynamic environments Complex, multi-objective tasks

Q&A

Exercises

Concepts

Define in your own words the following terms:

  • Rationality
  • Autonomy
  • Agent
  • Environment
  • Sensor
  • Actuator
  • Percept
  • Agent function
  • Agent program

Agent types

Explain the differences between the following agent types in your own words. Describe the component(s) that is/are specific for each type.

  • Reflex agent
  • Model-based agent
  • Goal-based agent
  • Utility-based agent
  • Learning agent

Vacuum cleaner

Under which circumstances does a robotic vacuum cleaner act rational?

Describe the task environment of such an agent.

PEAS

For each of the following agents, specify the performance measure, the environment, the actuators, and the sensors.

  • Microwave oven
  • Chess program
  • Autonomous supply delivery

Performance measure

Describe a task environment in which the performance measure is easy to specify completely and correctly, and a in which it is not.

Assertions

For each of the following assertions, say whether it is true or false and support your answer with examples or counterexamples where appropriate.

  1. An agent that senses only partial information about the state cannot be perfectly rational.
  2. There exist task environments in which no pure reflex agent can behave rationally.
  3. There exists a task environment in which every agent is rational.
  4. Every agent is rational in an unobservable environment.
  5. A perfectly rational poker-playing agent never loses.

Task environment

For each of the following activities characterize the task environment it in terms of the properties discussed in the lecture notes.

  • Playing soccer
  • Exploring the subsurface oceans of Titan
  • Shopping for used AI books on the internet
  • Playing a tennis match

Task environment #2

For each of the following task environment properties, rank the example task environments from most to least according to how well the environment satisfies the property.

Lay out any assumptions you make to reach your conclusions.

  1. Fully observable: driving; document classification; tutoring a student in calculus; skin cancer diagnosis from images
  2. Continuous: driving; spoken conversation; written conversation; climate engineering by stratospheric aerosol injection
  3. Stochastic: driving; sudoku; poker; soccer
  4. Static: chat room; checkers; tax planning; tennis

Literature

Acharya, Deepak Bhaskar, Karthigeyan Kuppan, and B Divya. 2025. “Agentic AI: Autonomous Intelligence for Complex Goals–a Comprehensive Survey.” IEEE Access.
Anthrophic. 2024. “Building Effective Agents.” Anthropic Research Team; https://www.anthropic.com/engineering/building-effective-agents.
Russel, Stuart, and Peter Norvig. 2022. Artificial Intelligence: A Modern Approach. Harlow: Pearson Education.
Wiener, Norbert. 1960. “Some Moral and Technical Consequences of Automation.” Science 131 (3410): 1355–58.

Footnotes

  1. The agent function maps any given percept sequence to an action (an abstract mathematical description).

  2. The term percept refers to the content an agent’s sensors are perceiving. The percept sequence is the complete history of everything an agent has ever perceived.

  3. Rectangles are used to denote the current internal state of the agent’s decision process, rectangles with rounded corners to represent the background information used in the process.