Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

chore: add function calling template for llama 3.1 models #3010

Merged
merged 1 commit into from
Jul 25, 2024

Conversation

mudler
Copy link
Owner

@mudler mudler commented Jul 25, 2024

Description

This PR adds function call capabilities to llama3.1 models by following https://github.com/meta-llama/llama-agentic-system/blob/ced0661761fc6529b23ac44ba1f19968ac5ad376/llama_agentic_system/system_prompt.py#L64

The function are returned in this syntax, for instance to call get_weather:

<function=get_current_weather>{{"location": "Boston, MA", "unit": "celsius"}}</function>

Copy link

netlify bot commented Jul 25, 2024

Deploy Preview for localai failed.

Name Link
🔨 Latest commit 16befcc
🔍 Latest deploy log https://app.netlify.com/sites/localai/deploys/66a28a3cfa1dd1000858fafa

Copy link

yamllint Failed

Show Output
::group::gallery/llama3.1-instruct.yaml
::error file=gallery/llama3.1-instruct.yaml,line=32,col=1::32:1 [trailing-spaces] trailing spaces
::endgroup::

Workflow: Yamllint GitHub Actions, Action: __karancode_yamllint-github-action, Lint: gallery

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
@mudler mudler merged commit 43f4953 into master Jul 25, 2024
20 of 31 checks passed
@mudler mudler deleted the llama3.1-functioncall branch July 25, 2024 17:37
truecharts-admin added a commit to truecharts/charts that referenced this pull request Jul 28, 2024
…9.3 by renovate (#24494)

This PR contains the following updates:

| Package | Update | Change |
|---|---|---|
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
patch | `v2.19.2-aio-cpu` -> `v2.19.3-aio-cpu` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
patch | `v2.19.2-aio-gpu-nvidia-cuda-11` ->
`v2.19.3-aio-gpu-nvidia-cuda-11` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
patch | `v2.19.2-aio-gpu-nvidia-cuda-12` ->
`v2.19.3-aio-gpu-nvidia-cuda-12` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
patch | `v2.19.2-cublas-cuda11-ffmpeg-core` ->
`v2.19.3-cublas-cuda11-ffmpeg-core` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
patch | `v2.19.2-cublas-cuda11-core` -> `v2.19.3-cublas-cuda11-core` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
patch | `v2.19.2-cublas-cuda12-ffmpeg-core` ->
`v2.19.3-cublas-cuda12-ffmpeg-core` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
patch | `v2.19.2-cublas-cuda12-core` -> `v2.19.3-cublas-cuda12-core` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
patch | `v2.19.2-ffmpeg-core` -> `v2.19.3-ffmpeg-core` |
| [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) |
patch | `v2.19.2` -> `v2.19.3` |

---

> [!WARNING]
> Some dependencies could not be looked up. Check the Dependency
Dashboard for more information.

---

### Release Notes

<details>
<summary>mudler/LocalAI (docker.io/localai/localai)</summary>

###
[`v2.19.3`](https://togithub.com/mudler/LocalAI/releases/tag/v2.19.3)

[Compare
Source](https://togithub.com/mudler/LocalAI/compare/v2.19.2...v2.19.3)

<!-- Release notes generated using configuration in .github/release.yml
at master -->

##### What's Changed

##### Bug fixes 🐛

- fix(gallery): do not attempt to delete duplicate files by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3031
- fix(gallery): do clear out errors once displayed by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3033

##### Exciting New Features 🎉

- feat(grammar): add llama3.1 schema by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3015

##### 🧠 Models

- models(gallery): add llama3.1-claude by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3005
- models(gallery): add darkidol llama3.1 by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3008
- models(gallery): add gemmoy by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3009
- chore: add function calling template for llama 3.1 models by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3010
- chore: models(gallery): ⬆️ update checksum by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#3013
- models(gallery): add mistral-nemo by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3019
- models(gallery): add llama3.1-8b-fireplace2 by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3018
- models(gallery): add lumimaid-v0.2-12b by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3020
- models(gallery): add darkidol-llama-3.1-8b-instruct-1.1-uncensored-iq…
by [@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3021
- models(gallery): add meta-llama-3.1-8b-instruct-abliterated by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3022
- models(gallery): add llama-3.1-70b-japanese-instruct-2407 by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3023
- models(gallery): add llama-3.1-8b-instruct-fei-v1-uncensored by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3024
- models(gallery): add openbuddy-llama3.1-8b-v22.1-131k by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3025
- models(gallery): add lumimaid-8b by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3026
- models(gallery): add llama3 with enforced functioncall with grammars
by [@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3027
- chore(model-gallery): ⬆️ update checksum by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#3036

##### 👒 Dependencies

- chore: ⬆️ Update ggerganov/llama.cpp by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#3003
- chore: ⬆️ Update ggerganov/llama.cpp by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#3012
- chore: ⬆️ Update ggerganov/llama.cpp by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#3016
- chore: ⬆️ Update ggerganov/llama.cpp by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#3030
- chore: ⬆️ Update ggerganov/whisper.cpp by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#3029
- chore: ⬆️ Update ggerganov/llama.cpp by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#3034

##### Other Changes

- docs: ⬆️ update docs version mudler/LocalAI by
[@&#8203;localai-bot](https://togithub.com/localai-bot) in
[mudler/LocalAI#3002
- refactor: break down json grammar parser in different files by
[@&#8203;mudler](https://togithub.com/mudler) in
[mudler/LocalAI#3004
- fix: PR title tag for checksum checker script workflow by
[@&#8203;dave-gray101](https://togithub.com/dave-gray101) in
[mudler/LocalAI#3014

**Full Changelog**:
mudler/LocalAI@v2.19.2...v2.19.3

</details>

---

### Configuration

📅 **Schedule**: Branch creation - At any time (no schedule defined),
Automerge - At any time (no schedule defined).

🚦 **Automerge**: Enabled.

♻ **Rebasing**: Whenever PR becomes conflicted, or you tick the
rebase/retry checkbox.

🔕 **Ignore**: Close this PR and you won't be reminded about these
updates again.

---

- [ ] <!-- rebase-check -->If you want to rebase/retry this PR, check
this box

---

This PR has been generated by [Renovate
Bot](https://togithub.com/renovatebot/renovate).

<!--renovate-debug:eyJjcmVhdGVkSW5WZXIiOiIzOC44LjMiLCJ1cGRhdGVkSW5WZXIiOiIzOC44LjMiLCJ0YXJnZXRCcmFuY2giOiJtYXN0ZXIiLCJsYWJlbHMiOlsiYXV0b21lcmdlIiwidXBkYXRlL2RvY2tlci9nZW5lcmFsL25vbi1tYWpvciJdfQ==-->
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant