Why Chat Mode is Dying: The Rise of Agentic Workflow

Agent LLM Flow Engineering Workflow

“Prompt Engineering” is dead, long live “Flow Engineering”?

Over the past year, we have witnessed an explosion in the capabilities of Large Language Models (LLMs). However, interacting with AI solely through a “Chat” window is rapidly hitting a ceiling in terms of utility. For writing poems, generating simple code, or answering encyclopedic questions, Zero-shot prompting works reasonably well. But when faced with complex real-world tasks—such as building a full-stack Web application, conducting deep market research, or debugging complex codebases—a single Prompt often falls short.

This is why we need to shift from Chat to Agentic Workflow.

The Bridge from Probability to Determinism

Fundamentally, an LLM is a probabilistic prediction machine. It doesn’t know the “correct answer”; it only knows the “next most probable token”. This probabilistic nature is an asset for creative generation, but when it comes to tasks requiring rigorous logic, the probability of getting a perfect result in a single shot is vanishingly small.

The core idea of an Agentic Workflow is not to expect the model to “get it right the first time”, but to design a closed-loop system that allows the model to “make mistakes and correct them”.

Core Patterns: Reflection and Tool Use

To make AI truly intelligent, we need to introduce two key capabilities:

  1. Reflection:
    It’s not just about generating content, but also generating an “evaluation” of that content. Just as a human developer writes a draft and then reviews it themselves. For example, a Coding Agent should not just output code, but follow a loop: Write Code -> Run Test -> Read Error -> Refine Code. This self-correction mechanism significantly improves the quality of the final output.

  2. Tool Use:
    Hallucination is a persistent issue with LLMs. Instead of asking the model to answer “what is the stock price right now” from memory, give it a get_stock_price() tool. When a model learns to “admit it doesn’t know, but knows where to look”, it evolves from a “knowledgeable parrot” into a “rigorous researcher”.

The Future: Flow Engineering

The future of software engineering may no longer be about writing specific business logic code, but about writing “Cognitive Flow”.

We need to define:

  • Who should the Agent turn to when stuck? (Web Search or Knowledge Base?)
  • What is the standard for completing a task? (Passing unit tests or User Review?)
  • How do multiple Agents (Manager, Coder, Reviewer) collaborate and hand off tasks?

This work of designing the interaction logic between prompts can be called Flow Engineering. Rather than trying to “hypnotize” the model with complex prompts, a well-designed Workflow can often achieve results with a smaller model that outperform larger models (GPT-3.5 + Good Workflow > GPT-4 Zero-shot).

Conclusion

Stop trying to find that “perfect Prompt”. Go build a system that tolerates errors, can reflect, and uses tools. This is the inevitable path to AGI.


Translated by Gridea Pro MCP from a Deep Dive post.

评论