Frameworks, Protocols, and Practical Tips for Building AI Agents

Peek into the practical side of AI agent development. We cover key frameworks like LangGraph, essential protocols like the new MCP and A2A, and tips to help developers build effective AI agents.

HIFLYLABS
|
|

Table of Contents

 

  • Introduction to AI agents
  • Core principles for development
  • Evolving frameworks
  • Let’s talk a bit about Model Context Protocol
  • So, how do MCP and A2A fit together?
  • What’s next?
  • Predictions on the operating principles of enterprise AI
  • Closing thoughts

 

 


 


You've seen it, right? LLMs got incredibly good, incredibly fast. They are now at the level of an intern dev, making plenty of mistakes, but at least very enthusiastically.

 

That rapid progress naturally makes us ask: what’s next? And how do we bring out the most of what we have today?

 

The continuing maturation of "AI agents" is one trend we can be sure of.

 

AI systems that don't just talk but take actions in the digital world use LLMs as their brain.

 

Usually, we see three key functions:

  • Planning: Figuring out the steps to reach a goal.
  • Tool Use: Interacting with other software, APIs, or data.
  • Reflection: Looking at their own work and adjusting things if needed.
agent_illu_02.jpg
One of the simplest ways of implementation: an agentic workflow.

 

Still, the current tech has its challenges, like LLM context limits and the occasional hallucination.

 

Though there are ways to manage them, you often need use case specific workarounds.

 

In this post, we want to cut through the hype and give you a practical overview of where AI agent development is right now—plus some predictions for the near future. We'll look at the core ideas, the frameworks we use, and some tools that might change how you build things.

 

Core principles for development

 

For the broad strokes of how you should effectively build agentic systems, Anthropic’s blog post is probably the gold standard. Key points:

 

1. Distinction between agentic workflows and full agents

  • Design: 
    • Workflows follow predefined paths. 
    • Agents direct their own path, and use tools based on environment feedback.
  • Use case: 
    • Workflow for well-defined tasks. 
    • Agents for open ended problem

 

2. Principles

  • Simplicity in design. (Especially when you first start building agents, use just an API call instead of a full agent.)
  • Prioritize transparency by explicitly showing the agent’s planning steps.
  • Carefully craft your agent-computer interface (ACI) through thorough tool documentation and testing.

 

3. Customization to use case

  • Combine and customize workflows
  • Customize model capabilities in retrieval, tools, and memory

 

These points resonate with the earlier overview of our own agent development approach, which is very accessible even for non-technical audiences. (Give it a read if you need to cover more basics.)

 

So we’ve covered how to think about building agents, but which actual tools and concepts do we use to put them into practice?

 

Evolving frameworks

 

Now, AI agents didn't appear overnight. Here are some highlights from recent years that made an impact on our daily work:

 

agent_illu_04.jpg

 

We first had foundational concepts like ReAct and Tree of Thoughts establish how the building blocks of agents should collaborate. These were put in practice, operationalized through the help of software frameworks like LangChain and LlamaIndex.

 

Initial excitement around early agents, like AutoGPT, exposed challenges. This led to us focusing on more structured, controllable systems.

 

There's also a clear trend towards frameworks offering explicit control over execution logic and state, moving from simple chains to graph-based approaches and multi-agent orchestration.

 

That brings us to LangGraph, made specifically for building stateful agents using controllable graph workflows. For us, it quickly became a go-to for building agents in 2024, and still is today. A huge plus is its checkpointing system. Every step, every state the agent goes through in the graph, can be saved. You can see exactly how many tokens were used, what decisions were made. This makes monitoring way easier. You don't have to build a custom logging system from scratch. You can save this state in memory, or in a SQL database, and easily pull out all that info after the agent has run. It gives you incredible visibility into what the agent actually did (see the demo towards the end of our latest webinar). Of course, it’s far from the only option if you’re looking for these capabilities, and there are other players who specialize in different interaction patterns.

 

Initially, we eagerly followed the progress that seemingly took parallel tracks: flexible open-source frameworks, vs. integrated platform solutions of big labs. Luckily, there is now a stronger initiative for interoperability and standardization in context management and tool integration.

 

agent_illu_01.jpg
Usual building blocks of an AI agent system today.

 

OpenAI has just released their own 34 page long practical guide to building agents. (See page 17 for multi-agent systems.) This white paper is great for starting out, however, there are actually some who question its contents.

 

Harrison Chase, the CEO of Langchain, took it deeply to heart when OpenAI referred to their own Agents SDK 'framework' as non-declarative, subtly criticizing more declarative frameworks like LangGraph or LlamaIndex Workflows to be inflexible, complex, hard to maintain—which is misleading.

 

In reality, the mentioned frameworks enable developers to clearly define high-level structure while also having control over behavior and logic precisely by adopting declarative and imperative approaches.

 

This is missing in OpenAI's Agents SDK—or Google's for that matter. They are both sets of abstractions. Great to quickly put together demos, spin up tool-calling agents, but they do not have robust orchestration layers. Simply defining handoffs and tool calls will yield an AI app with less human control (high variance) and 'fake' flexibility.

 

Chase was motivated enough to start comparing agent frameworks, and put together a spreadsheet, (which will hopefully be updated from time to time).

 

With all that said, the trend is clear: 

 

Agentic systems nowadays are much more reliable and easier to build, despite the growing number of capabilities and complexity of available options and sources. How is that possible?

 

Let’s talk a bit about Model Context Protocol

 

Building an agent's "brain" is one thing. The trick was always getting them to reliably interact with APIs, databases, files, other tools, and other AI agents.

 

Historically, every time you wanted your agent to use a new tool or data source, you had to build a custom integration. Which could have scaling or reliability issues, or both.

 

Enter Model Context Protocol.

 

Anthropic kicked off MCP in late 2024, aiming to create a universal standard for connecting LLMs to anything. Not just their own models, but anyone’s.

 

How’s it different from tool calling? Well, it works consistently across different models and platforms. 

 

Whether it's talking to a Git repo, a SQL database, or a web API, MCP aims for minimal setup.

 

It basically works like this:

  • Host Application: Your app running the LLM.
  • MCP Client: Translates requests between your app and the tool server.
  • MCP Server: Exposes the tool's capabilities (accessing APIs, files, etc.).
  • Transport Layer: Handles the communication (locally or over HTTP).

 

Why is this a big deal for developers?

 

It makes adding tools to your agents way easier, especially for frameworks like LangChain or even ReAct-style agents.

 

Define your tool server once. Hook it into your custom code, whatever. No more reinventing the wheel for every integration.

 

But MCP only really started gaining steam when OpenAI embraced the standard in March 2025 as well, along with their SDK and Responses API releases.

 

The community jumped on it. We quickly saw collections like the awesome-mcp-servers repo pop up, linking agents to hundreds of real tools.

 

Just last week, LangChain even launched an open-source library to connect any LLM to MCP tools easily. 

 

Then, Google entered the chat with their own standard agent dev kit (ADK) and Agent-to-Agent Protocol (A2A). Things are moving fast!

 

A2A focuses specifically on how multiple agents can collaborate securely and effectively, especially across different platforms or when tasks take a long time.

 

So, how do MCP and A2A fit together?

 

MCP acts as the standard for an agent (the "Host") to talk to its tools (via MCP Servers). A2A, on the other hand, is designed for agents to talk to each other. It adds features MCP doesn't focus on, like secure authentication between agents, managing shared tasks and state, and discovering what other agents can do.

 

They're complementary. You might have multiple agents (communicating via A2A) where each agent uses MCP to access its own set of tools. This distinction is quite important.

 

It's still early days, and we might see some competition as both protocols evolve. But the overall trend towards standardization is great news for us devs trying to build more complex, interconnected systems.

 

What’s next?

 

So, these tools and frameworks are great incremental steps. But what's the bigger picture?

 

Many expect the next leap to come from inference scaling, aka allowing models to run much longer, more complex planning and action sequences.

 

Agents would be one step closer to automating entire workflows, not just small parts of tasks.

 

There's this idea floating around: a new Moore's Law for AI agents. The length of a task an AI can complete autonomously is doubling every 7 months or so.

 

ai-digest-new-moores-law-agent-task-completion-time.png
“A new Moore's Law for AI agents: When ChatGPT came out in 2022, it could do 30 second coding tasks. Today, AI agents can autonomously do coding tasks that take humans an hour.” – AI task completion ability seems to be growing exponentially. 
Source: https://theaidigest.org/time-horizons

 

Today, agents can tackle tasks that take humans an hour. If trends hold, in a few years, they might handle tasks that take us a month. That's a path towards seriously powerful capabilities.

 

What does this mean for us developers (besides a lot of reward hacking)?

 

Will AI make our jobs obsolete? We don't think so, at least not in the next 5-6 years.

 

Instead, we foresee a shift:

  • Our job descriptions will change. Generator vs validator problem.)
  • There will likely be a shortage of talent with the right AI-centric skills.
  • Development processes and rhythms will transform.
  • Traditional programming languages might become increasingly abstracted away by natural language interfaces.

 

The generator vs validator problem won’t go away for some time (AI generates, humans validate the results). Even though we already see o3 being able to do almost 50% of OpenAI’s internal pull requests (see figure 20 of their System Card).

 

But the idea of "vibe coding"—prompting without deep technical understanding—probably won't scale very well in the short term. 

 

Enterprise-grade applications still need skilled humans to:

  • Ask the right questions
  • Add necessary considerations and criteria
  • Implement design patterns that amateurs would never ever know to prompt for 
  • Make judgment calls about tradeoffs

 

Until we reach superintelligence or the literal technological singularity, it's still humans augmenting AI for almost all real-world uses.

 

But even if we go with the most conservative and grounded scenario of AI evolution, the trend that coding assistants themselves are becoming increasingly autonomous and capable is undeniable. Tools like GitHub Copilot, Cursor, Replit, Firebase Studio or Codex CLI can tackle bigger tasks across your whole project, suggesting edits, analyzing errors, etc… helping generate full-stack apps from prompts in record time, if you know what you’re doing.

 

You could say that yeah, AI assistance in your IDEs is great for speeding up things. But deploying full-stack applications is a bit more involved. Indeed, it is.

 

We will be releasing a writeup—coming next month—with specific best practices and architecture blueprints for every step of the development lifecycle (code review, bug fixes, testing, etc). This white paper will be sourced directly from hands-on experience from our most recent, real-world client projects. If you really don’t want to miss it, follow us on Linkedin.

 

Beyond the immediate impact on developer workflows and tools, deploying agents at scale across an organization suggests a fundamental shift in how enterprises might operate. Thinking about that future, here are some potential operating principles that seem to be emerging:

 

Predictions on the operating principles of enterprise AI

 

Some takeaways from our recent projects and the overall trends discussed above:

  • Orchestration will shift from rule-based to context-based: dynamic, adaptive, truly intelligent. Agent networks will decide who leads based on intent, not hierarchy. The monolith is dead. Long live the mesh of intelligent agents.
  • Agents are not products. They're participants in workflows. Salesforce, Workday, etc... each will own its workflow, but not the full customer journey. Legacy IT is already struggling. Agent-based architectures will widen the gap. Successful adoption will be key to staying in the game. 
  • At large, AI adoption won't fail because of model capabilities. It'll fail because of a lack of interoperability and integration. The battle for AI supremacy will be won in orchestration, not inference.
  • MCPs and Agent-to-Agent standards will become the TCP/IP of enterprise AI. Agent-to-agent coordination is the enterprise glue of the AI era. AI is no longer a layer; it's the fabric stitching the enterprise together.
  • AgentOps will become the new DevOps. Composable AI is like Lego for workflows. You bring your blocks. The system will build itself.
  • You won't debug code, you'll debug conversations between agents. Building an agent is easy. Getting 50 to work together is not. Agent interactions will need the same auditability and traceability as financial systems. How your agents reason will be subject to compliance. An agent is only as trustworthy as the data it's trained on.
  • Enterprise AI will be a choreography of agents across your stack, not centralized. But enterprise IT isn't ready. Most data isn't even accessible, let alone AI-ready. Agent networks will force a reckoning with your data infrastructure. You'll need governance not just over data, but over agent behavior.
  • Vertical agents will dominate first. Horizontal orchestration will follow.

 

Closing thoughts

 

So, what's the bottom line?

 

A clear direction towards simplification.

 

Building basic agents, especially ones that call tools, is definitely getting easier.

 

As the underlying models get smarter, they also need less hand-holding.

 

Expect less reliance on rigidly defined workflows. We’ll only be needed for very high level agent design.

 

The goal is for agents to deeply understand what you need and figure out how to get there themselves. That's the exciting part.

Article by HIFLYLABS
Artificial Intelligence
AI Agents

Explore more stories

Flying high with Hifly

We want to work with you

Hiflylabs is your partner in building your future. Share your ideas and let’s work together.