Skip to content

Latest commit

 

History

History
91 lines (60 loc) · 2.67 KB

README.md

File metadata and controls

91 lines (60 loc) · 2.67 KB

Function Calling With Modus

LLM apis such as OpenAI, have a feature called function calling or tool use. With this feature the LLM response to a chat message could be a request to invoke a function, usually in order to collect information necessary to generate a response.

This project demonstrates how to setup function calling within Modus, the open source framework for building intelligent APIs.

The example implements a function askQuestionToWarehouse accepting an query in natural language about prices or stock of goods in the warehouse.

The API uses 2 tools available for the LLM

  • get_product_types: provide the list of product types we have in the warehouse
  • get_product_info: return an info (qty or price) about one product type

Get started

1- Set your credentials

Create the file .env.dev.local in api-as folder, containing your OpenAI API key:

MODUS_OPENAI_API_KEY="sk-...."

2- launch the API

From api-as folder launch

modus dev

3- Test the GraphQL operation From a GraphQL client (Postman), Introspect the GraphQL endpoint http://localhost:8686/graphql Invoke the operation askQuestionToWarehouse

# example query using tool calling

query AskQuestion {
  askQuestionToWarehouse(question: "What is the most expensive product?") {
    response
    logs
  }
}

The operation returns the final response and an array of strings showing showing the tool calls and messages exchanged with the LLM API.

Experiment with some queries to see the function calling at work.

# example of questions
What can you do for me?
what fo we have in the warehouse?
How many shoes  in stock?
How many shoes and hats do we have in stock?
what is the price of a desks?
What is the most expensive product in stock?

Details

The logic is as follow:

  • Instruct the LLM to use function calls (tools) with the correct parameters to get the data necessary to reply to the provided question.

  • Execute the identified function calls in Modus to build an additional context (tool messages)

  • Re-invoke the LLM API with the additional tool messages.

Return the generated responses based on the data retrieved by the function calls.

Discussion

Correct prompt helps to address questions that are out of scope.

Descriptions of function and parameters are also part of the prompt engineering!

Enums parameter can help. Try replacing the product_name parameter by an Enum type and see that the LLM can skip a function call.

Need a way to avoid loops. That's why we have a limit to 3 calls.

Need to experiment more to understand what are good functions in terms of abstraction, number of parameters etc ...