As AI systems continue to evolve from passive responders to autonomous agents capable of planning, reasoning, and executing tasks independently, we stand at a critical inflection point. The rise of agentic AI — models that can reason, plan, and act towards goals with minimal human instruction — presents enormous potential. Yet it also demands a fresh rethinking of what it means to build trustworthy and responsible AI.
Trust must be designed in from the start. When it comes to agentic AI, embedding principles of responsibility, alignment, and transparency is essential to ensure that autonomy enhances, rather than erodes, the human experience. Below are seven guiding principles inspired by global best practices and leading AI governance frameworks such as the OECD AI Principles, the EU AI Act, Singapore’s Model AI Governance Framework, and NIST’s AI Risk Management Framework.

- Purpose-Aligned Agency
Autonomous agents must pursue goals that are clear, ethical, and aligned with human and societal intent. That starts with defining the agent’s mandate — not merely what it can do, but what it should do. This is closely aligned with the OECD AI Principles of human-centred values and fairness. We advocate for constraint-based autonomy: systems that are flexible within defined ethical, legal, and operational boundaries.
- Transparent Planning and Reasoning
As agents plan and reason more independently, the ability to explain their decisions becomes essential. This includes:
- Traceability of actions and planning logic
- Logging of decisions, thresholds, and underlying assumptions
- Human-readable pathways that foster trust among both users and auditors
These measures echo the NIST AI Risk Management Framework’s emphasis on explainability and interpretability. Transparency is not only a technical requirement; it’s a social contract.
- Human-in-the-Loop and Human-on-the-Loop
Autonomy must never come at the expense of human oversight. We distinguish between:
- Human-in-the-loop: required for real-time, critical decisions
- Human-on-the-loop: providing supervisory control, capable of intervening when needed
This is consistent with Singapore’s Model AI Governance Framework, which encourages clear roles and escalation procedures. Clear escalation protocols are key — particularly when agents encounter ambiguity, ethical dilemmas, or potential harm.
- Alignment with Legal and Ethical Norms
Agentic systems must remain anchored in ethical and legal frameworks. This includes:
- Embedding values through methods such as preference learning and value alignment
- Ensuring compliance with emerging AI regulations, such as the EU AI Act’s risk-based classification and obligations for high-risk AI
- Designing agents capable of self-monitoring for policy violations or grey areas
Responsibility is not just about compliance — it’s about long-term societal impact.
- Built-In Guardrails, Adaptive by Design
No system is infallible. We advocate for layered technical guardrails, including:
- Pre-deployment: rigorous scenario testing and stress-testing based on high-risk AI criteria under the EU AI Act
- Post-deployment: continuous behavioural monitoring and the use of watchdog agents
- Fail-safes and rate-limiters: to prevent runaway behaviour or cascading failures
This echoes the resilience goals of the NIST AI Risk Management Framework, which calls for robust, secure, and reliable AI systems.
- Trust Signals for Users and Stakeholders
People need to know when they are interacting with an agent, not a human — and what the system’s intent and limitations are. Key trust enablers include:

- Clear labelling of agentic interactions
- Communication of confidence levels and contextual caveats
- Mechanisms for users to question or challenge decisions
Trust is earned through clarity, not complexity. These ideas are supported by the transparency and accountability principles in the OECD and Singaporean frameworks.
- Continuous Learning, Morally Anchored
Agentic systems are not static. They learn and adapt. But that learning must be morally anchored.
We advocate for:
- Ethical memory: preserving lessons learned over time
- Bounded learning: avoiding drift into unaligned or unsafe behaviours
- Human-AI co-learning: where agents adapt through shared experience, not isolation
This supports principles of sustainability and accountability, ensuring that learning systems remain aligned over time.
A Mnemonic to Remember: The TRUST-AGENT Framework
To aid both design and dialogue, we summarise this approach in a simple mnemonic:
- T – Transparency & Traceability
- R – Responsible Reasoning
- U – User Oversight
- S – Safety & Scenario Testing
- T – Trust Signals
Final Thoughts
Agentic AI represents a new era — one where machines do more than respond; they reason, decide, and act. But without intentional design, this autonomy risks eroding the very trust that underpins its adoption. Our collective responsibility is to ensure that as AI becomes more agentic, it also becomes more human-aligned — not less.
Because in the end, the most powerful agents aren’t the ones that act independently, but the ones that act responsibly.


Leave a comment