Singapore Institute of Technology

A one-day course ↯

Agentic AI
for Applied Learning.

Learn agent principles, then put them to work in your teaching practice.

principles

Our checklist for today

  • Default to curiosity and make your thinking tangible
    Use the chatbot in the corner to ask questions and park your ideas.
  • Bias toward action
    Don't overthink. Try it, test it, and learn by doing.
  • Step into your students' shoes
    Notice what feels confusing, exciting, or overwhelming as you learn — it's what they'll feel too.
today's journey

From n00b to Agentic Educator.

n00b Agentic Educator
Foundations
What agents actually are
Anatomy
Inside an agent
Design
Agentic design patterns
Evaluations
Evaluate your agent
First Build
Your first agent
Workshop Session
Solve for a real client
↯ click the cards ↯ why you're here today

The Agentic Opportunity

"

For every SaaS company, there will be an AI agent company.

- Harj Taggar, Partner Y Combinator

flip to reveal ↗

Agents democratize economic opportunities.

"

Fast experimentation drives invention.

- Andrew Ng, Computer Scientist, former head of Google Brain & Chief Scientist at Baidu

flip to reveal ↗

The key enabler of AI is fast development and iteration, and that allows for more experimentation.

proof point · real-world deployment

SingTel Case Study

0 wks
Went live in under 10 weeks with Sierra's forward-deployed engineer (FDE)
0+
Customer cases handled in the first six weeks
0%
Mobile and home troubleshooting resolved without a human agent
0%
Roaming sign-ups completed without human intervention
freed
up
Customer care officers freed up to focus on complex, higher-value work
↯ click to explore ↯ the definition

What is an agent?

A model that perceives inputs, reasons about what to do, takes actions - and keeps going, step by step, until the goal is met.

GOAL PERCEIVE text · files · data · web REASON plan what to do next ACT use tools · take steps ITERATE check · retry · adapt
Core elements
Brain The LLM core

The LLM enables the agent's reasoning. You swap models to change its reasoning engine. The agent is not its model.

Sonnet ChatGPT Gemini Spark Deepseek Qwen
Memory What the agent can remember

By default an agent remembers your conversation up to a limit and forgets anything mentioned earlier. You can give it more memory like documents, past conversations or research papers and it will behave like a personal librarian.

The trade-off: the more it remembers, the more it costs to run and the easier it loses focus.

Tools How the agent acts

Without tools the agent can only chat. Tools can let it search the web, read files, write spreadsheets. Think of tools as the agent's hands - they give it the ability to do things.

Autonomy How much it acts on its own

Some agents ask for permission at every step while others just run until the job is done. The right level of autonomy depends on what's at stake. You wouldn't let an agent send emails to 10,000 students without a human checking the draft first.

Iteration How it works through a problem

It doesn't just answer once and stop. It tries something, checks if it worked, and decides what to do next, looping until the goal is met. This is what separates an agent from a chatbot. A chatbot responds. An agent persists.

foundations

A non-agentic workflow is a recipe.
An agent is a cook.

Non-agentic / Automation

predictable
  • One-pass execution
  • Predefined, fixed steps
  • No revisions, no iteration
  • "If X → do Y" rules
  • think: zero/one-shot workflows like summarising PDFs, rewriting emails, generating images
vs

Agentic

adaptive
  • ReasoningReasons about what to do next
  • MemoryRemembers past interactions
  • Access to toolsUses tools to act on the world
  • IterationIterates until goal is met
  • think: a digital employee
⟳ drag to turn ⟳ from prompt to autopilot

Autonomy is a dial, not a switch.

ZERO ORONE-SHOT HUMAN INTHE LOOP AGENTIC FULLYAUTONOMOUS

Use case
    Industries
      making the call

      Decision Guide: Agent or Automation?

      Question 1

      Can you map out every step the system should take, in advance?

      Yes ↙ ↘ No
      Question 2a

      Does the work mainly involve interpreting messy human content?

      Question 2b

      Would quality improve if the system critiques its own draft?

      Yes ↙ ↘ No Yes ↙ ↘ No
      Recommendation

      AI step in a workflow

      Known pipeline with one step that reads unstructured text.

      Recommendation

      Traditional automation

      Predictable steps on structured data — a script beats an LLM.

      Recommendation

      Reflective agent

      Open-ended path where self-critique improves the output.

      Recommendation

      Tool-using agent

      Open-ended path — the agent picks tools and acts on the world.

      designing agents

      Agent Design Canvas.

      Before you build, fill in the text for this design canvas to form a complete brief for your agent. Flip each card to see how the six elements work together in this example of a billing support assistant that reads customer emails and decides what to do next.

      flip each card to fill the canvas
      01Role
      Who - or what - is this agent?
      flip to reveal ↓
      Name the job title, domain, and authority level. Be precise - the model fills every gap you leave.
      e.g. Billing support specialist. Handles inbound subscription billing disputes and account queries. Acts as first-line customer support.
      Skip it → Scope expands unpredictably. The model defaults to agreeable and overpromising.
      02Task
      What is it trying to accomplish?
      flip to reveal ↓
      State one output. Name the input, the transformation required, and what a correct result looks like. One task per agent.
      e.g. Read an inbound billing email. Classify the issue, set urgency 1-3, draft a reply, and flag if a human needs to step in.
      Skip it → The agent self-defines its goal. It optimises for something - just not necessarily what you intended.
      03Context
      What does it know going in?
      flip to reveal ↓
      List every data source by name: what's in the prompt, what it retrieves via tools, and what state it carries between turns.
      e.g. The customer's email, their account history from Salesforce, the refund policy doc, and their last 3 support tickets.
      Skip it → Data gaps get filled with confident-sounding invention. The agent won't flag what it doesn't know.
      04Tools
      What can it connect to?
      flip to reveal ↓
      Name every system it can look things up in or take action in. State what it can look up versus what it can change.
      e.g. Look up records in Salesforce, search the help centre, create tickets in Zendesk. Cannot edit billing records directly.
      Skip it → It only knows what it was trained on. No live systems, no actions.
      05Guardrails
      What must it never do?
      flip to reveal ↓
      Write explicit rules: what it must never do, what requires human approval before acting, and what triggers escalation.
      e.g. Never promise a refund without checking account status. Escalate if refund > $200, legal is mentioned, or 3+ open tickets.
      Skip it → The agent acts on best judgement. One edge case triggers an irreversible action.
      06Output
      What should the result look like?
      flip to reveal ↓
      Describe exactly what it should hand back - the format, what each part contains, and what a correct result looks like.
      e.g. Issue type, urgency (1-3), draft reply, escalate (yes/no), and reason if escalating. No free-form summaries.
      Skip it → The agent decides the format itself. If it varies, anything reading it downstream breaks.
      canvas complete - ready to build
      how agents are structured

      Agentic Design Patterns.

      / 04

        Best for

        Examples
        ways to organise agents

        Multi-agent Systems.

        / 04

        When to reach for it

        ↯ watch it refine ↯ measuring agent quality

        Evaluations: Measuring Agent Accuracy & Effectiveness.

        inputs agent outputs score 80% 100% 04 / 05 passing · fix the failing case 05 / 05 passing · regression fixed
        catch → fix → re-run

        How to run Evals

        1. 01
          Pick 10 real examples. Actual requests from real users - the inputs your agent has to handle day to day.
        2. 02
          Write down the expected result. For each one, note the answer or action you'd want back. That's your ground truth.
        3. 03
          Run and score the output. Compare each result to your ground truth - turn it into a percentage using a metric below.
        4. 04
          Refine and re-run. Tweak the prompt, swap the tool, change the flow - re-run the same examples and watch the score move.
        5. 05
          Grow the sample set over time. Every new failure becomes the next example. The broader the set, the more real the score.

        What to measure

        • Task success passes / total × 100%

          e.g. support agent resolves a refund request end-to-end

        • Trajectory sensible paths / total × 100%

          e.g. research agent actually reads the papers, not just titles

        • Tool accuracy correct tool calls / total × 100%

          e.g. sales agent queries the right CRM field with the right filters

        • Cost & time avg. $ & seconds per run

          e.g. coding assistant holds $0.12 / 8s avg, not drifting to $0.40 / 30s

        • Human-intervention rescued runs / total × 100%

          e.g. how often a human rep has to take over a support conversation

        ↯ sort · filter ↯ the agent stack

        Tools for building agents.

        Use
        Build
        Lift
        License
        Tool What it is Pick it for… No-code Lift
        Claude / CoworkGeneral Strong reasoning assistant plus an agentic desktop that works across your files and apps. Knowledge workers who need deeper reasoning, polished writing, or a desktop agent across files and apps. Yes Low–Med
        Gemini + WorkspaceGeneral AI built into Gmail, Docs, Meet, plus Gemini app and NotebookLM through Workspace plans. Orgs already in Google - meeting notes, docs, slides, and classroom prep in one plan. Yes Low
        Perplexity ComputerGeneral Cloud-hosted agentic task runner that coordinates 19 AI models to browse the web, control apps, fill forms, and complete multi-step workflows from natural language. Business users who want to delegate browser-based research and multi-step workflows in plain English - no setup, runs entirely in the cloud. Yes Low
        NotebookLMGeneral Source-grounded research and learning tool that works on your uploaded materials. Grounded learning and teaching prep - synthesise, query, and audio-overview your own sources. Yes Low
        Zapier AgentsNo-code No-code agents connected to thousands of apps, designed to act on business data and triggers. Quick no-code automations connecting business apps - the fastest path from idea to running agent. Yes Low
        GumloopNo-code Visual, AI-native agent builder where you drag-and-drop nodes (130+ integrations, MCP support) to create reasoning agents and multi-step automations, with a meta-agent that scaffolds flows from plain-language prompts. Teams wanting AI-native automation over simple triggers - agents that reason mid-flow, not just chain apps; step up from Zapier when logic and branching matter. Yes Low–Med
        n8nNo-code Workflow automation for technical teams with AI agents, traceable reasoning, visual builder, and self-hosting. Technical teams bridging workflow automation and real agent systems, with self-hosting options. Yes +code Medium
        Relevance AINo-code Low/no-code platform for building AI agents and multi-agent teams, especially around GTM and operations. Sales, onboarding, and GTM teams building AI agent workflows without writing code. Yes Medium
        FlowiseNo-code Open-source drag-and-drop builder for LLM agents and RAG apps with 100+ model and vector-store integrations, self-host or managed cloud. Developers and small teams prototyping chat and RAG agents visually, with the option to self-host or embed via API. Yes Low–Med
        DifyNo-code Open-source LLMOps platform combining agentic workflows, RAG pipelines, tools, and observability - no-code builder with code escape hatches. Teams building production AI apps end-to-end - citizen developers get a visual builder, engineers get APIs, evals, and self-hosting. Yes +code Low–Med
        MindStudioNo-code No-code visual builder for designing, deploying, and managing AI agents and apps across 200+ models and 1,000+ integrations - publishable as web apps, APIs, browser extensions, or scheduled automations. Teams and agencies building user-facing AI products without engineers - step above prompt wrappers, with 150K+ deployed agents and an approachable visual interface. Yes Low
        OpenAI CodexDeveloper Autonomous coding agent - available as web UI, desktop app, open-source CLI, and IDE extensions - that writes features, fixes bugs, runs tests, and orchestrates parallel sub-agents in sandboxed cloud environments. Engineering teams wanting an agent that writes, tests, and ships code autonomously - hands-off software engineer, accessible via CLI, API, or ChatGPT web UI. Partial Medium
        LangChainDeveloper Developer framework and platform (LangGraph + LangSmith) for building, testing, observing, and deploying custom agents. Developers building bespoke agent systems with full control over evals, memory, and observability. No High
        SierraEnterprise Enterprise customer-experience AI platform with agent studio, SDK, live assist, voice, and trust tooling. Enterprise CX teams deploying branded, voice-ready customer agents at scale. Platform-led Med–High
        OpenHandsOpen-source Open platform for cloud coding agents. Formerly OpenDevin; now stewarded by All Hands AI. Software engineering agents and sandboxed dev workflows - the leading open-source coding agent. No High
        OpenClawOpen-source Open-source self-hosted personal AI assistant with tool use, plugins, browser/canvas/actions, and companion apps. Power users building a personal agent layer with plugins, browser actions, and self-hosting. Partial Med–High
        Hermes AgentOpen-source Self-improving open-source agent by Nous Research with persistent memory, reusable skills, and a learning loop. Long-running autonomous agents that need cross-session memory and self-improving skills. Partial Med–High
        No tools match your filters.
        agent tooling checklist

        How to pick a tool for building agents.

        01 Lift

        How quickly you can go from signing up to a working agent.

        AskCould someone without a coding background build their first agent in an afternoon?

        Red flagThe "getting started" guide asks you to install software on your computer before you can try anything.

        02 Model choice

        Swap the AI brain without rebuilding the agent.

        AskCan I switch from Claude to GPT to a cheaper model with a single setting?

        Red flagYou're locked to one AI company — switching means starting over.

        03 Connections

        A big library of apps and tools it works with, and a way to add your own.

        AskCan I connect the agent to something that isn't on the default list?

        Red flagOnly the apps they've pre-built work. Your internal tools aren't welcome.

        04 Visibility

        The ability to see exactly what the agent did, step by step.

        AskWhen the agent gets something wrong, can I quickly see why?

        Red flagWhen it breaks, you get an error message and no trail back to the cause.

        05 Guardrails

        Permissions, sensitive data handling, approvals, and a clear record.

        AskCan I require a human to approve before the agent does anything risky or permanent?

        Red flagAnyone logged in can do anything, and nothing is recorded.

        06 Room to grow

        The thing you demo is the thing you ship to real users.

        AskWhen real usage picks up, can this tool help me scale — or do I rebuild somewhere else?

        Red flagGoing live means rebuilding the whole agent in a different system.

        hands-on build · dify

        Building your first reflection agent.

        What this agent does

        the idea
        • InputAn essay, an assignment brief, and a grading rubric.
        • ScoreReads the essay and scores it against your rubric (1–10).
        • BranchIf score ≥ 7, stops and returns the assessment.
        • ReviseOtherwise, identifies the top weaknesses and rewrites the essay to address them.
        • LoopRepeats until the essay passes, or up to 3 iterations.
        • you get the final essay, the score, and how many tries it took

        How to get started

        • 1Go to dify.aiGet Started
        • 2Sign up with email, Google, or GitHub.
        • 3Open the Studio tab, then Workflow.
        • 4Create from Blank → select Workflow.
        • 5App name: Iterative Essay Agent, then Create.
        • then open this Google doc and follow the rest →
        agent design studio

        Real client. Real problem.
        Forty minutes. One good agent.

        Your rules of engagement

        • Get into groups.
        • Pick one problem worth solving for the business. Use real data from the business.
        • Plan your agent on the canvas before you touch Dify.
        • Use your preferred AI to build a working proof of concept in Dify.
        • You have 40 minutes for the above!
        • Present to the client. Five minutes per group.

        What you'll need to define for your agent

        • Agent goal (Use the Agent Design Canvas)
        • Data sources (structured vs unstructured)
        • Human-in-the-loop checkpoints (if any)
        • Pick one evaluation metric to assess the quality of your agent
        workshop

        Workshop: Agent Design Canvas.

        Fill in each card to form the brief for the agent you want to build.

        0 / 6 fields filled
        01Role
        Who - or what - is this agent?
        02Task
        What is it trying to accomplish?
        03Context
        What does it know going in?
        04Tools
        What can it connect to?
        05Guardrails
        What must it never do?
        06Output
        What should the result look like?
        to close

        Tools change,
        knowledge compounds.

        The meta-skill is learning to evaluate, switch, and compose tools - not mastering any one. What you built today is the muscle for everything that comes next.

        thanks for coming ↯
        doodle mode on · press D or esc to exit