DLite V2: Lightweight, Open LLMs That Can Run Anywhere
<p>AI Squared is committed to democratizing AI so that it can be used by all. There are two key forces opposing the democratization of AI though — a tendency for high-performing models to have a huge number of parameters, making them incredibly expensive to train, tune, and deploy at scale — and nonpermissive licensing preventing many open source models from being used for commercial purposes.</p>
<p><img alt="" src="https://miro.medium.com/v2/resize:fit:560/0*_DXI1ZU5v2FDR88E" style="height:281px; width:700px" /></p>
<p><em>Getting high performance from smaller models would greatly reduce the start-up and operational costs of building with large language models.</em></p>
<p>To address the size / cost aspect of this current situation, in April 2023 , we released our DLite V1 family of models, which are lightweight LLMs ranging from 124 million parameters to 1.5 billion parameters that exhibit ChatGPT-like interactivity. The small size of these models means that they can be run on almost any device, including laptop CPUs, instead of being limited to deployment on specialized, expensive cloud resources. At this point, however, we were using the Alpaca dataset to tune the model, which prevented any of the DLite v1 family from being used for commercial purposes.</p>
<p>We’ve since updated the DLite family with DLite V2, which also has four different models ranging from 124 million to 1.5 billion parameters. The highlight of the update was our utilization of the `databricks-dolly-15k` dataset released by Databricks. We have also uploaded this dataset to our HuggingFace page so anyone can easily use it. Because this training dataset is also licensed for commercial purposes, we are also happy to announce that all models in the DLite V2 family can also be used for commercial purposes, enabling organizations to build upon these models with no licensing constraints whatsoever.</p>
<p><a href="https://medium.com/ai-squared/dlite-v2-lightweight-open-llms-that-can-run-anywhere-3ab59b3a7feb">Website</a></p>