Workflow Group Chat


This guide demonstrates how to build and run a multi-agent group chat workflow using the AIGNE Framework. You will learn how to orchestrate several agents, including a manager, to collaborate on a task, simulating a team environment where they share messages and work together to achieve a common goal.

Overview#

The Group Chat workflow example showcases a sophisticated multi-agent system where different agents with specialized roles collaborate to fulfill a user's request. The process is managed by a Group Manager agent that directs the conversation and task execution among other agents like a Writer, Editor, and Illustrator.

This example supports two primary modes of operation:

  • One-shot mode: The workflow runs once to completion based on a single input.
  • Interactive mode: The workflow engages in a continuous conversation, allowing for follow-up questions and dynamic interactions.

The core interaction model is as follows:


Prerequisites#

Before proceeding, ensure your development environment meets the following requirements:

  • Node.js: Version 20.0 or higher.
  • npm: Included with Node.js.
  • OpenAI API Key: Required for the default model configuration. You can obtain one from the OpenAI Platform.

Quick Start#

You can run this example directly without cloning the repository using npx.

Run the Example#

Execute one of the following commands in your terminal:

To run the workflow in the default one-shot mode:

Run in one-shot mode

npx -y @aigne/example-workflow-group-chat

To start an interactive chat session:

Run in interactive mode

npx -y @aigne/example-workflow-group-chat --chat

You can also provide input directly via a pipeline:

Run with pipeline input

echo "Write a short story about space exploration" | npx -y @aigne/example-workflow-group-chat

Connect to an AI Model#

The first time you run the example, it will prompt you to connect to an AI model provider since no API keys have been configured.

Initial setup prompt for connecting to an AI model.

You have several options to proceed:

1. Connect to the AIGNE Hub (Recommended)#

This is the easiest way to get started and includes free credits for new users.

  1. Select the first option: Connect to the Arcblock official AIGNE Hub.
  2. Your web browser will open a page to authorize the AIGNE CLI.
  3. Click "Approve" to grant the necessary permissions. The CLI will be configured automatically.

Authorization dialog for AIGNE Hub connection.

2. Connect to a Self-Hosted AIGNE Hub#

If you are running your own instance of AIGNE Hub:

  1. Select the second option: Connect to your self-hosted AIGNE Hub.
  2. Enter the URL of your AIGNE Hub instance when prompted.
  3. Follow the instructions in your browser to complete the connection.

Prompt to enter the URL for a self-hosted AIGNE Hub.

3. Configure a Third-Party Model Provider#

You can directly connect to a provider like OpenAI by setting an environment variable.

  1. Exit the interactive prompt.
  2. Set the OPENAI_API_KEY environment variable in your terminal:

    Configure OpenAI API Key

    export OPENAI_API_KEY="your-openai-api-key"
  3. Run the example command again.

For other providers like Google Gemini or DeepSeek, refer to the .env.local.example file within the project for the correct environment variable names.

Local Installation and Usage#

For development purposes, you can clone the repository and run the example locally.

1. Clone the Repository#

Clone the framework repository

git clone https://github.com/AIGNE-io/aigne-framework

2. Install Dependencies#

Navigate to the example's directory and install the required packages using pnpm.

Install dependencies

cd aigne-framework/examples/workflow-group-chat
pnpm install

3. Run the Example#

Use the pnpm start command to run the workflow. Command-line arguments must be passed after --.

To run in one-shot mode:

Run in one-shot mode

pnpm start

To run in interactive chat mode:

Run in interactive mode

pnpm start -- --chat

To use pipeline input:

Run with pipeline input

echo "Write a short story about space exploration" | pnpm start

Command-Line Options#

The example accepts several command-line arguments to customize its behavior:

Parameter

Description

Default

--chat

Run in interactive chat mode

Disabled (one-shot mode)

--model <provider[:model]>

AI model to use in format 'provider[:model]' where model is optional. Examples: 'openai' or 'openai:gpt-4o-mini'

openai

--temperature <value>

Temperature for model generation

Provider default

--top-p <value>

Top-p sampling value

Provider default

--presence-penalty <value>

Presence penalty value

Provider default

--frequency-penalty <value>

Frequency penalty value

Provider default

--log-level <level>

Set logging level (ERROR, WARN, INFO, DEBUG, TRACE)

INFO

--input, -i <input>

Specify input directly

None

Examples#

Set logging level

pnpm start -- --log-level DEBUG

Use a specific model

pnpm start -- --model openai:gpt-4o-mini

Debugging with AIGNE Observe#

To inspect the execution flow and debug the behavior of the agents, you can use the aigne observe command. This tool launches a local web server that provides a detailed view of agent traces.

First, start the observability server in a separate terminal:

Start the observability server

aigne observe

Terminal output showing the AIGNE Observe server starting.

After running the workflow example, open your browser to http://localhost:7893 to view the traces. You can inspect the inputs, outputs, and internal states of each agent throughout the execution.

AIGNE Observe web interface showing a list of traces.

Summary#

This guide provided a step-by-step walkthrough for running the Workflow Group Chat example. You learned how to execute the workflow using npx, connect to various AI model providers, and install it locally for development. You also saw how to use aigne observe for debugging agent interactions.

For more complex patterns, explore other examples in the AIGNE Framework documentation.