# NexusRaven-13B **Repository Path**: hf-models/NexusRaven-13B ## Basic Information - **Project Name**: NexusRaven-13B - **Description**: 一个用于日期处理任务的模型。 - **Primary Language**: Unknown - **License**: Not specified - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 1 - **Forks**: 0 - **Created**: 2023-10-25 - **Last Updated**: 2024-11-28 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README --- license: llama2 base_model: codellama/CodeLlama-13b-Instruct-hf model-index: - name: NexusRaven-13B results: [] --- # NexusRaven-13B: Surpassing the state-of-the-art in open-source function calling LLMs.
Nexusflow HF - NexusRaven blog post - NexusRaven-13B - NexusRaven-13B Twitter Thread - NexusRaven-13B Github - NexusRaven API evaluation dataset
Table of contents - [NexusRaven-13B: Surpassing the state-of-the-art in open-source function calling LLMs.](#nexusraven-13b-surpassing-the-state-of-the-art-in-open-source-function-calling-llms) - [Introducing NexusRaven-13B](#introducing-nexusraven-13b) - [NexusRaven model usage](#nexusraven-model-usage) - [Training procedure](#training-procedure) - [Training hyperparameters](#training-hyperparameters) - [Framework versions](#framework-versions) - [Limitations](#limitations) - [License](#license) - [References](#references) - [Citation](#citation) - [Contact](#contact) This model is a fine-tuned version of [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf). ## Introducing NexusRaven-13B NexusRaven is an open-source and commercially viable function calling LLM that surpasses the state-of-the-art in function calling capabilities. 📊 Performance Highlights: With our demonstration retrieval system, NexusRaven-13B achieves a 95% success rate in using cybersecurity tools such as CVE/CPE Search and VirusTotal, while prompting GPT-4 achieves 64%. It has significantly lower cost and faster inference speed compared to GPT-4. 🔧 Generalization to the Unseen: NexusRaven-13B generalizes to tools never seen during model training, achieving a success rate comparable with GPT-3.5 in zero-shot setting, significantly outperforming all other open-source LLMs of similar sizes. 🔥 Commercially Permissive: The training of NexusRaven-13B does not involve any data generated by proprietary LLMs such as GPT-4. You have full control of the model when deployed in commercial applications. ## NexusRaven model usage NexusRaven accepts a list of python functions. These python functions can do anything (including sending GET/POST requests to external APIs!). The two requirements include the python function signature and the appropriate docstring to generate the function call. NexusRaven is highly compatible with langchain. See [langchain_example.py](https://huggingface.co/Nexusflow/NexusRaven-13B/blob/main/langchain_example.py). An example without langchain can be found in [non_langchain_example.py](https://huggingface.co/Nexusflow/NexusRaven-13B/blob/main/non_langchain_example.py). Please note that the model will reflect on the answer sometimes, so we highly recommend stopping the model generation at a stopping criteria of `["\nReflection:"]`, to avoid spending unnecessary tokens during inference, but the reflection might help in some rare cases. This is reflected in our langchain example. More information about how to prompt the model can be found in [prompting_readme.md](prompting_readme.md). The "Initial Answer" can be executed to run the function. ### Quickstart You can run the model on a GPU using the following code. ```python # Please `pip install transformers accelerate` from transformers import pipeline pipeline = pipeline( "text-generation", model="Nexusflow/NexusRaven-13B", torch_dtype="auto", device_map="auto", ) prompt_template = """