-
Win Gifts from Second State/WasmEdge at KubeCon+CloudNativeCon+ OSSummit+AI_dev 2024
The GenAI, cloud-native and pen source community is eagerly anticipating the upcoming KubeCon + CloudNativeCon + Open Source Summit + AI_dev China 2024, set to take place in Hong Kong 21-23rd August. This event promises to be a remarkable gathering of open-source luminaries, including the legendary Linus Trovalds and other star speakers. Developers will have the rare opportunity for face-to-face interactions with these influential figures, as well as with Jim Zemlin, the CEO of the Linux Foundation, and other industry leaders.…
-
Advance Your Skills with WasmEdge LFX Mentorship 2024 Fall: LLMs, Trading Bots and More
The 2024 Term 3 of the LFX Mentorship program is here, and it's packed with exciting opportunities! Running from September to November, this program invites passionate developers to contribute to open source while boosting their CV. WasmEdge has 4 projects that offer an exciting opportunity for aspiring developers to work on cutting-edge projects within the WasmEdge ecosystem, with a focus on enhancing WebAssembly (WASM) capabilities, improving software reliability, and integrating modern AI techniques to build innovative applications.…
-
Mathstral: A New LLM that is Good at Math Reasoning
Today, Mistral AI released mathstral, a finetuned 7B model specifically designed for math reasoning and scientific discovery. The model has a 32k context window. The model weights are available under the Apache 2.0 license. As we have seen, leading edge LLMs, such as the GPT-4o, can solve very complex math problems. But do they have common sense? A meme that has been going around on the Internet suggests that LLMs can only pretend to solve “math Olympiad level” problems since it lacks understanding of even elementary school math.…
-
Getting Started with internlm2_5-7b-chat
The internlm2_5-7b-chat model, a new open-source model from SenseTime, introduces a 7 billion parameter base model alongside a chat model designed for practical applications. This model showcases exceptional reasoning capabilities, achieving state-of-the-art results in math reasoning tasks, outperforming competitors like Llama3 and Gemma2-9B. With a remarkable 1M context window, InternLM2.5 excels in processing extensive data, leading in long-context challenges such as LongBench. The model is also capable of tool use, integrating information from over 100 web sources, with enhanced functionalities in instruction adherence, tool selection, and reflective processes.…
-
Building a Translation Agent on LlamaEdge
By MileyFu, CNCF Ambassador, DevRel and Founding Member of WasmEdge runtime. Prof. Andrew Ng's agentic translation is a great demonstration on how to cooridnate multiple LLM “agents” to work on a single task. It allows multiple smaller LLMs (like Llama-3 or Gemma-2) to work gether and produce better results than a single large LLM (like ChatGPT). The translation agent is a great fit for LlamaEdge, which provides a lightweight, embeddable, portable, and Docker-native AI runtime for many different types of models and hardware accelerators.…
-
Getting Started with Gemma-2-9B
Google recently released Gemma 2 models in 9B and 27B Sizes, which are the latest models of its Gemma models family. According to its technical report, there will be an open sourced Gemma-2-2b model in the upcoming days. The technical report also demonstrates that the Gemma-2-9B model outperforms the Mistral-7B, Llama-3-8B, and the Gemma 1.5 models in several benchmarks. In this article, taking Gemma-2-9B as an example, we will cover…