Quantcast
Channel: Machine Learning | Towards AI
Viewing all articles
Browse latest Browse all 792

Snowflake‘s New LLM is Open Source and Built for the Enterprise

$
0
0
Author(s): Jorge Alcántara Barroso Originally published on Towards AI. Snowflake’s New Arctic LLM is Open Source and Built for Enterprise 𝐖𝐡𝐚𝐭 𝐀𝐈 𝐖𝐢𝐧𝐭𝐞𝐫? ❄ A new chill model is breaking the ice this spring: Snowflake’s Arctic is served mixed, cool, and open-source – and it is coming for your Enterprise use cases. The cost of entry for large-scale AI development has traditionally been a major roadblock, limiting advanced capabilities to deep-pocketed tech giants. Snowflake’s recent unveiling of the Arctic model is poised to disrupt this landscape. Powered by an innovative Mixture of Experts (MoE) architecture and released under the permissive Apache 2.0 license, Arctic puts both cost-effective scalability and open-source adaptability into the hands of businesses. This isn’t just about a single language model. Snowflake is following Databrinks in releasing powerful, open-source models targetting enterprise use, definitely signaling where the industry is headed. Arctic’s focus on performance for business applications, along with its computational efficiency, opens up the possibility of custom-tailored AI solutions without breaking the bank. Let’s take a good look into what makes Arctic unique, and explore how this new model and MoE architectures will impact enterprise AI, what are its current limitations, and where we expect to see the area grow in the future. Why is Open Source important? Closed source models have so far maintained a clear lead on performance. OpenAI and Anthropic currently offer the definite model winners: GPT-4 and Claude 3 Opus, only accessible through their (and hosting partners) APIs. Check out the latest benchmark results at HuggingFace, rankings at 🏆 LMSYS Chatbot Arena Leaderboard, and usage comparisons at OpenRouter. However, the majority of applications do not require the highest-performing models, and as these models spend more time under the sun, we have seen a clear shift toward fine-tuned task-specific models that perform better in their narrow use cases. The open-source projects have proven time and again to be the catalysts for innovation. When the source code and model weights are freely available — as with Arctic — a much wider community of developers and businesses can examine, modify, and build upon the work. Unlike proprietary “black box” models, open-source AI offers several key advantages: Customization: Businesses have the freedom to fine-tune the model specifically for their own data and use cases, leading to potentially higher accuracy and a more seamless integration into existing workflows. Rapid Experimentation: With no licensing hurdles, developers can rapidly experiment with different configurations and approaches, accelerating the development of novel use cases. Community-Driven Progress: Open-source encourages collaboration. Businesses, researchers, and enthusiast developers can pool their expertise, potentially leading to even faster improvements than those achievable within a single company. Source: The State of Enterprise Open Source: A Red Hat report Licensing Differences: Arctic vs. Llama 3 Not all open source is created equal. Understanding the licensing terms is crucial for businesses and developers aiming to leverage these technologies. The licensing framework not only impacts the usability of the models but also defines the boundaries within which companies can operate and innovate. Llama 3’s License stipulates several conditions for use, reproduction, distribution, and modification. Some key elements: Limited License Grant: Users are granted a non-exclusive, worldwide, non-transferable, and royalty-free license to use and modify the Llama Materials, though this comes with certain restrictions, particularly in redistribution. Attribution Requirements: Redistribution of Llama Materials or derivative works requires inclusion of the original agreement and a clear attribution to Meta Llama 3, such as “Built with Meta Llama 3.” Commercial Use Limitations: If the monthly active users of a product using Llama 3 exceed 700 million, a separate license must be negotiated with Meta, potentially introducing a barrier for large-scale deployments. No Warranty and Limitation of Liability: The license explicitly states that the Llama Materials are provided “AS IS” without any warranties and limits Meta’s liability, shifting risk onto the user. In contrast, Snowflake’s Arctic model utilizes the Apache 2.0 license, which is much more permissive, particularly in the aspects that matter to our clients: Permissive Use: Allows users to use, modify, and distribute the software with fewer restrictions. There is no requirement for a user agreement or active user limits. Redistribution: of original or derivative works does not require the inclusion of the original license or explicit attribution to the original software, meaning flexibility in how the software is presented and marketed. Commercial Use: No restrictions based on the scale of use, making it ideal for both startups and large enterprises to deploy at scale without needing to renegotiate terms or incur additional licensing fees. There are tangible implications for businesses aiming to integrate these AI models into their products and services. For instance, a company looking to develop a proprietary software solution that incorporates an LLM could utilize Arctic without the need to attribute the technology to Snowflake or reveal the use of open-source software in its marketing materials. Unpacking the MoE Architecture At its core, a Mixture of Experts (MoE) model departs from the traditional approach to language models where a single, massive network of parameters handles every task. Think of the dense architecture as attempting to have a single doctor who is equally skilled in cardiology, dermatology, and neurology. While possible, there are limits to such a generalist’s capabilities. Instead, MoE is like a specialist hospital. It houses a collection of smaller, highly specialized “expert” networks. For each input (a line of text, a SQL query), a “routing” mechanism decides which experts are most relevant. Only those selected experts get activated, significantly reducing the computational overhead compared to activating a giant, dense model in its entirety. This means: Greater Scale at Lower Cost: MoE models can achieve impressive performance and handle massive datasets with training costs scaling better than dense solutions. Specialization: The modular nature of MoE allows for fine-grained specialization. Consider an expert focused purely on legal code interpretation, while another might excel at understanding scientific terminology. Source: Scaling Laws for Fine-Grained Mixture of Experts The power of MoE truly shines in its ability to deliver exceptional […]

Viewing all articles
Browse latest Browse all 792

Trending Articles