Skip to main content

AI Agent Action: Token Usage Optimization

AI Agent actions in workflows now use fewer tokens per execution without reducing output quality.

Who This Is For / When to Use

Use this update to:

  • Reduce AI usage costs across workflows

  • Scale AI-driven automations more efficiently

  • Improve performance of high-volume workflows

  • Run longer or more complex AI processes within the same budget

What Changed

The AI Agent action has been optimized to reduce unnecessary token usage at every stage of execution.

Cleaner Context Handling

  • Removed duplicate and unnecessary data from prompts

  • Excludes raw database internals and redundant contact fields

  • Reduces payload size sent to the model

Smarter Tool Responses

  • Tool outputs now include only relevant fields

  • Example: Instead of dozens of fields, only key data like:

    • Name

    • Email

    • Phone

  • Eliminates excess data being passed into the model

Conversation Memory Optimization

  • Older conversation steps are automatically summarized

  • Recent messages remain in full detail

  • Prevents sending full conversation history on every execution

Structured Output Improvements

  • Final outputs no longer duplicate full context

  • Reduces token-heavy extraction steps

How It Works

The system automatically optimizes token usage during AI Agent execution.

  1. Workflow triggers the AI Agent action

  2. System filters and compresses context data

  3. Tool responses are minimized to essential fields

  4. Older conversation data is summarized

  5. AI generates output using optimized input

Performance Improvements

Measured on identical workflows with no loss in output quality:

  • First LLM call: ~36% fewer tokens

  • Total execution: ~20% fewer tokens

Why This Matters

Every AI Agent execution now:

  • Uses fewer tokens

  • Costs less to run

  • Maintains the same response quality

This allows:

  • More workflow executions within the same budget

  • Better scalability for AI-powered automation

  • Improved efficiency across all AI use cases

Common Issues and Fixes

Output looks shorter than before

  • This is expected due to removal of redundant data

  • Core response quality and accuracy are unchanged

Concern about missing context

  • Relevant context is still preserved

  • Older data is summarized, not removed

Token usage still high

  • Review workflow design (large inputs, long prompts, or many actions can still increase usage)

FAQ

Does this affect AI response quality?

No. Output quality remains the same. Only redundant or unnecessary data has been removed.

Do I need to enable this optimization?

No. This improvement is automatic for all AI Agent actions.

Are older conversations still used?

Yes. Older interactions are summarized while recent ones are kept in full detail.

Will this reduce my AI costs?

Yes. Since billing is based on tokens, reduced usage directly lowers cost.

Can I control how much context is sent?

Not directly. The system automatically determines the most efficient context size.

Did this answer your question?