Skip to content

docs: added a data analysis between model size over tool calling performance #931

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 2 commits into
base: main
Choose a base branch
from

Conversation

anyuzoey
Copy link

@anyuzoey anyuzoey commented Mar 3, 2025

Summary

This PR adds an analysis of how model size correlates with various performance metrics in the Berkeley Tool-Calling Function Leaderboard.

Key Updates

Preprocessed leaderboard data (cleaning, filtering, and structuring).
Filtering out proprietary models (as most of those doesn't publish their model size.)
manual edit model size column and summarise as a csv file.
Correlation analysis between model size and accuracy features.
Visualisation with a heatmap to highlight trends.

Motivation

As part of my personal research, I aim to understand how model size impacts performance across different metrics and identify patterns in scaling. I feel this analysis is also useful for general users and thats why i raise a PR here.

anyuzoey added a commit to anyuzoey/awesome-small-language-model-1B-3B that referenced this pull request Mar 5, 2025
manually adding a model size column to their leaderboard for analysing model size versus performance. added this to gorilla: ShishirPatil/gorilla#931
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant