<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Function-Calling | Zhang Handuo's Site</title><link>https://handuo.top/tags/function-calling/</link><atom:link href="https://handuo.top/tags/function-calling/index.xml" rel="self" type="application/rss+xml"/><description>Function-Calling</description><generator>Hugo Blox Builder (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Sat, 18 Jan 2025 00:00:00 +0000</lastBuildDate><image><url>https://handuo.top/media/icon_huc589ef2b3ca02823d26b34234e0ca591_160546_512x512_fill_lanczos_center_3.png</url><title>Function-Calling</title><link>https://handuo.top/tags/function-calling/</link></image><item><title>FC-Model</title><link>https://handuo.top/project/watt_fc/</link><pubDate>Sat, 18 Jan 2025 00:00:00 +0000</pubDate><guid>https://handuo.top/project/watt_fc/</guid><description>&lt;ul>
&lt;li>&lt;a href="https://gorilla.cs.berkeley.edu/leaderboard/">Leaderboard&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://huggingface.co/watt-ai/watt-tool-70B">Huggging Face Model Card 70B&lt;/a>&lt;/li>
&lt;li>&lt;a href="https://huggingface.co/watt-ai/watt-tool-8B">Hugging Face Model Card 8B&lt;/a>&lt;/li>
&lt;/ul>
&lt;p>We opensourced a 70B (and a 8B) model for function calling and tool usage. The model is based on LLaMa-3.3-70B-Instruct, optimized for tool usage and multi-turn dialogue. It achieves state-of-the-art performance on the Berkeley Function-Calling Leaderboard (BFCL).&lt;/p>
&lt;h2 id="model-description">Model Description&lt;/h2>
&lt;p>This model is specifically designed to excel at complex tool usage scenarios that require multi-turn interactions, making it ideal for empowering platforms like &lt;a href="https://lupan.watt.chat">Lupan&lt;/a>, an AI-powered workflow building tool. By leveraging a carefully curated and optimized dataset, watt-tool-70B demonstrates superior capabilities in understanding user requests, selecting appropriate tools, and effectively utilizing them across multiple turns of conversation.&lt;/p>
&lt;p>Target Application: AI Workflow Building as in &lt;a href="https://lupan.watt.chat/">https://lupan.watt.chat/&lt;/a> and &lt;a href="https://www.coze.com/">Coze&lt;/a>.&lt;/p>
&lt;h2 id="key-features">Key Features&lt;/h2>
&lt;ul>
&lt;li>&lt;strong>Enhanced Tool Usage:&lt;/strong> Fine-tuned for precise and efficient tool selection and execution.&lt;/li>
&lt;li>&lt;strong>Multi-Turn Dialogue:&lt;/strong> Optimized for maintaining context and effectively utilizing tools across multiple turns of conversation, enabling more complex task completion.&lt;/li>
&lt;li>&lt;strong>State-of-the-Art Performance:&lt;/strong> Achieves top performance on the BFCL, demonstrating its capabilities in function calling and tool usage.&lt;/li>
&lt;li>&lt;strong>Based on LLaMa-3.1-70B-Instruct:&lt;/strong> Inherits the strong language understanding and generation capabilities of the base model.&lt;/li>
&lt;/ul>
&lt;h2 id="training-methodology">Training Methodology&lt;/h2>
&lt;p>watt-tool-70B is trained using supervised fine-tuning on a specialized dataset designed for tool usage and multi-turn dialogue. We use CoT techniques to synthesize high-quality multi-turn dialogue data.&lt;/p>
&lt;p>The training process is inspired by the principles outlined in the paper: &lt;a href="https://arxiv.org/abs/2406.14868">&amp;ldquo;Direct Multi-Turn Preference Optimization for Language Agents&amp;rdquo;&lt;/a>.
We use SFT and DMPO to further enhance the model&amp;rsquo;s performance in multi-turn agent tasks.&lt;/p></description></item></channel></rss>