Functional Tokens for Creating Enterprise-Grade Agentic Systems

Functional tokens streamline enterprise-grade agentic systems by enhancing function prediction efficiency in language models.

Functional tokens are becoming a crucial component in developing enterprise-grade agentic systems. These tokens enable more efficient and accurate planning by transforming how functions are predicted and executed within language models. Today, we will delve into the methodology behind using functional tokens for agentic systems, highlighting the advantages and demonstrating a practical implementation.

Traditional Planning Methods and Their Challenges

Most companies engage in prompt engineering for planning tasks, where they provide the model with a set of functions and request it to generate a plan in a specific JSON format. This approach often includes Chain of Thought prompting to enhance the model’s reasoning capabilities. However, this method faces several challenges:

  1. Prompt Size and Hallucination: Adding numerous functions can make the prompt excessively large, leading to hallucinations and incorrect plans by the LLM.
  2. Data Control: Sending data to the cloud for complex planning poses risks and constraints, as smaller models are typically incapable of handling such tasks.
  3. Cost: Agentic systems require numerous API calls, resulting in high operational costs.

Introducing Octopus v2: A New Approach

The paper “Octopus v2: On-device language model for super agents” proposes an innovative technique to fine-tune models to predict functions directly. This approach involves fine-tuning the model to output functional tokens, simplifying the planning process. Here’s a brief overview of the methodology:

  1. Tokenizer Enhancement: Adding new tokens to the tokenizer for each function.
  2. Fine-Tuning: Training the model on a dataset to predict these tokens based on the input questions.
  3. Efficiency: This method eliminates the need for analyzing tokens from function descriptions, thereby avoiding the retrieval and processing of these descriptions.
  4. Model Used: The Gemma 2B model was used for fine-tuning in the Octopus v2 paper.

This approach allows the language model to treat function calling as a standard completion task, enhancing efficiency and accuracy.

Implementing Functional Tokens with Phi3 Model

I have fine-tuned a Phi3 model to handle single functional calls using a small synthetic dataset generated with ChatGPT. The fine-tuning process was completed using Unsloth AI. Below is a practical demonstration of how to implement this:

Step 1: Setup and Installation

Ensure you have the necessary libraries installed. If not, you can install them using pip:

Step 2: Tokenizer Enhancement

First, add new tokens to the tokenizer for each function:

Step 3: Fine-Tuning the Model

Fine-tune the model on a dataset to predict the functional tokens. Here is a simplified example:

Step 4: Using the Fine-Tuned Model

After fine-tuning, you can use the model to predict functions directly:

Conclusion

Using functional tokens significantly enhances the efficiency and accuracy of agentic systems. By fine-tuning models to predict these tokens directly, we can overcome the challenges of traditional prompt engineering methods, such as prompt size limitations, data control issues, and high costs. The methodology outlined in the Octopus v2 paper and demonstrated with the Phi3 model provides a practical approach to implementing functional tokens in enterprise-grade systems.

For more details and the full implementation, check out the GitHub repository and the Octopus v2 paper.

Picture of Association of Data Scientists

Association of Data Scientists

The Chartered Data Scientist Designation

Achieve the highest distinction in the data science profession.

Elevate Your Team's AI Skills with our Proven Training Programs

Strengthen Critical AI Skills with Trusted Generative AI Training by Association of Data Scientists.

Our Accreditations

Get global recognition for AI skills

Chartered Data Scientist (CDS™)

The highest distinction in the data science profession. Not just earn a charter, but use it as a designation.

Certified Data Scientist - Associate Level

Global recognition of data science skills at the beginner level.

Certified Generative AI Engineer

An upskilling-linked certification initiative designed to recognize talent in generative AI and large language models

Join thousands of members and receive all benefits.

Become Our Member

We offer both Individual & Institutional Membership.