🚀Calling all developers of copilots and AI agents! Introducing 🐦⬛NexusRaven V2, a 13B function calling LLM surpassing GPT-4 in real-world zero-shot tool use.
✨ Highlights of 🐦⬛NexusRaven V2:
💪Superior Performance: NexusRaven V2 surpasses GPT-4 up to 7% on complex nested and
🆕 Check out the recent update of 𝕎𝕚𝕝𝕕𝔹𝕖𝕟𝕔𝕙! We have included a few more models including DBRX-Instruct
@databricks
and StarlingLM-beta (7B)
@NexusflowX
which are both super powerful! DBRX-Instruct is indeed the best open LLM; Starling-LM 7B outperforms a lot of even
🔍 What Starling-LM-7B-beta's excellent performance tells us about benchmarks
I compared the performance of
@NexusflowX
's model across various benchmarks.
In the Chatbot Arena Leaderboard (), this 7B model impressively outperforms many larger models,
Apple MLX: considering the power of Starling-LM-7B-beta from
@NexusflowX
and its ranking on
@lmsysorg
Chatbot Arena Leaderboard, I converted and uploaded 4bit and 8bit versions on HuggingFace mlx-community!
Performance 🔥 on M2 Ultra 76GPU:
- 4bit: tokens/sec Prompt: 158 -
Have we really squeezed out the capacity of a compact chat model? Thrilled to see our latest open model, Starling-7B, ranks 13th among all models in Chatbot Arena!
🚀 As a 7B model, Starling surpasses larger open and proprietary models, including Claude-2, GPT-3.5-Turbo, Gemini
[Arena Update]
70K+ new Arena votes🗳️ are in!
Claude-3 Haiku has impressed all, even reaching GPT-4 level by our user preference! Its speed, capabilities & context length are unmatched now in the market🔥
Congrats
@AnthropicAI
on the incredible Claude-3 launch!
More exciting
🚀 Presenting Starling-LM-7B-beta, our cutting-edge 7B language model fine-tuned with RLHF!
🌟 Also introducing Starling-RM-34B, a Yi-34B-based reward model trained on our Nectar dataset, surpassing our previous 7B RM in all benchmarks.
✨ We've fine-tuned the latest Openchat
🚀 Presenting Starling-LM-7B-beta, our new cutting-edge 7B language model fine-tuned with RLHF!
🌟 Also introducing Starling-RM-34B, the workhorse Reward Model behind the Starling-LM-7B-beta, ranking
#1
in the latest RewardBenchmark from
@natolambert
and the
@allenai_org
team.
📢 Powerful information extraction app built by
@theanakin87
and
@Haystack_AI
team using
#NexusRaven
-V2 LLM for function calling!
🔥 We are thrilled to empower high-quality Gen AI apps with our compact LLMs and toolings.
🧪📑 Open LLMs with function calling capabilities can be used to extract data in a structured form
See my experiment with
@Haystack_AI
+ NexusRaven by
@NexusflowX
📓
🚀 Exciting breakthrough in LLM reliability! 🧠NexusRaven-V2, our cutting-edge function-calling LLM, has set a new standard in minimizing AI hallucinations, surpassing GPT-4's performance in a recent third-party independent research benchmark.
Dive into our latest blog post to
Thank you,
@bigdata
for extending your platform to us and engaging in fantastic conversation!🎙 Excited to contribute to the foundation of
#AI
Agents!
🎧LISTEN:
Thank you,
@deci_ai
and
@DataScienceHarp
, for featuring NexusRaven-V2 in the top 10 compact & robust models. Stay tuned for what's to come for the community in 2024!
Excited to share that NexusRaven-V2 was featured in
@AIatMeta
's Llama newsletter this morning!📩
If you’re new here, NexusRaven-V2 is a 13B function calling LLM surpassing GPT-4 in real-world zero-shoot tool use.
💻 Learn more about NexusRaven-V2’s capabilities:
Nexusflow team checking in from
#NeurIPS2023
.
Want to check out fun copilots using our technologies? Come to Booth 423 and hang out with folks behind NexusRaven-V2 () and Starling LLMs ().
On Friday, Dec. 15, we’ll present
The Nexusflow team heads to
#NeurIPS2023
next week! 📍
Find us at 𝗕𝗼𝗼𝘁𝗵 423 to get an exclusive demo application built on top of NexusRaven-V2 – our new model that surpasses GPT-4 in using tools for building agents and copilots.
On Friday, Dec. 15, we’ll also present