Microsoft adds GPT-4 Turbo LLM to free version of Copilot
The year 2023 saw Microsoft betting heavily on artificial intelligence and its partnership with OpenAI to make Copilot a reality.
With quantum LLMs now available on HuggingFace and AI ecosystems like H20, Text Gen, and GPT4All allowing you to load LLM weights on your computer, you now have an option for free, flexible, and secure AI. Here are the 9 best local/offline LLMs you can try right now!
Table of Contents

Hermes 2 Pro is a state-of-the-art language model fine-tuned by Nous Research. It uses an updated and compact version of the OpenHermes 2.5 dataset, along with the newly introduced Function Calling and JSON datasets developed by the company. The model is based on the Mistral 7B architecture and has been trained on 1,000,000 instructions/conversations of GPT-4 quality or better, mostly synthetic data.
|
Model |
Hermes 2 Pro GPTQ |
|---|---|
|
Model size |
7.26 GB |
|
Parameters |
7 billion |
|
Quantization |
4-bit |
|
Type |
Mistral |
|
License |
Apache 2.0 |
The Hermes 2 Pro on the Mistral 7B is the new flagship Hermes 7B model, offering improved performance across a variety of benchmarks, including AGIEval, BigBench Reasoning, GPT4All, and TruthfulQA. Its enhanced capabilities make it suitable for a wide range of natural language processing (NLP) tasks, such as code generation, content creation, and conversational AI applications.
Zephyr is a series of language models trained to act as helpful assistants. Zephyr-7B-Beta is the second model in the series, fine-tuned from Mistral-7B-v0.1 using Direct Preference Optimization (DPO) on a mix of publicly available synthetic datasets.
|
Model |
Zephyr 7B Beta |
|---|---|
|
Model size |
7.26 GB |
|
Parameters |
7 billion |
|
Quantization |
4-bit |
|
Type |
Mistral |
|
License |
Apache 2.0 |
By removing the built-in alignment of the training datasets, Zephyr-7B-Beta demonstrates improved performance on benchmarks like MT-Bench, increasing its usefulness for a variety of tasks. However, this adjustment can lead to problematic text generation when prompted in certain ways.

This quantized version of Falcon is based on a decoder-only architecture fine-tuned on TII's raw Falcon-7b model. The base Falcon model is trained using 1.5 trillion outstanding tokens sourced from the public Internet. As an Apache 2-licensed, command-based decoder-only model, Falcon Instruct is perfect for small businesses looking for a model to use for language translation and data ingestion.
|
Model |
Falcon-7B-Instruct |
|---|---|
|
Model size |
7.58 GB |
|
Parameters |
7 billion |
|
Quantization |
4-bit |
|
Type |
Falcon |
|
License |
Apache 2.0 |
However, this version of Falcon is not ideal for fine-tuning and is only intended for inference. If you want to fine-tune Falcon, you will need to use the raw model, which may require access to enterprise-grade training hardware like NVIDIA DGX or AMD Instinct AI Accelerators.
GPT4All-J Groovy is a decoder-only model tuned by Nomic AI and licensed under Apache 2.0. GPT4ALL-J Groovy is based on the original GPT-J model, which is known to be great at generating text from prompts. GPT4ALL-J Groovy has been tuned into a conversational model, which is great for fast and creative text generation applications. This makes GPT4All-J Groovy ideal for content creators in assisting them with their writing and composition, whether it is poetry, music, or stories.
|
Model |
GPT4ALL-J Groovy |
|---|---|
|
Model size |
3.53 GB |
|
Parameters |
7 billion |
|
Quantization |
4-bit |
|
Type |
GPT-J |
|
License |
Apache 2.0 |
Unfortunately, the baseline GPT-J model was trained on an English-only dataset, which means that even this fine-tuned GPT4ALL-J model can only converse and perform text generation applications in English.

DeepSeek Coder V2 is an advanced language model that enhances programming and mathematical reasoning. DeepSeek Coder V2 supports multiple programming languages and provides extended context length, making it a versatile tool for developers.
|
Model |
DeepSeek Coder V2 Instruct |
|---|---|
|
Model size |
13 GB |
|
Parameters |
33 billion |
|
Quantization |
4-bit |
|
Type |
DeepSeek |
|
License |
Apache 2.0 |
Compared to its predecessor, DeepSeek Coder V2 shows significant improvements in coding, reasoning, and general performance. It expands support for programming languages from 86 to 338 and extends the context length from 16K to 128K tokens. In benchmarks, it outperforms models such as GPT-4 Turbo, Claude 3 Opus, and Gemini 1.5 Pro in cryptographic and mathematical benchmarks.

Mixtral-8x7B is a mixture of expert (MoE) model developed by Mistral AI. It has 8 experts per MLP, totaling 45 billion parameters. However, only two experts are activated per token during inference, making it computationally efficient, with speed and cost comparable to a 12 billion parameter model.
|
Model |
Mixtral-8x7B |
|---|---|
|
Model size |
12 GB |
|
Parameters |
45 billion (8 experts) |
|
Quantization |
4-bit |
|
Type |
Mistral MoE |
|
License |
Apache 2.0 |
Mixtral supports context lengths of 32k tokens and outperforms Llama 2 by 70B on most benchmarks, matching or exceeding GPT-3.5 performance. It is fluent in multiple languages, including English, French, German, Spanish, and Italian, making it a versatile choice for a variety of NLP tasks.
Wizard-Vicuna GPTQ is the quantum version of Wizard Vicuna based on the LlaMA model. Unlike most LLMs released to the public, Wizard-Vicuna is an uncensored model with de-linking. This means that the model does not have the same safety and ethical standards as most other models.
|
Model |
Wizard-Vicuna-30B-Uncensored-GPTQ |
|---|---|
|
Model size |
16.94 GB |
|
Parameters |
30 billion |
|
Quantization |
4-bit |
|
Type |
LlaMA |
|
License |
GPL 3 |
While it can pose a problem to control AI alignment, having an uncensored LLM also brings out the best in the model by allowing it to respond without any constraints. This also allows users to add their own custom alignment to how the AI should act or respond based on a given prompt.

Looking to test a model trained using a unique learning approach? Orca Mini is an informal implementation of Microsoft’s Orca research papers. The model is trained using a teacher-student learning approach, where the dataset is filled with explanations rather than just prompts and feedback. This should theoretically make the student smarter, as the model can understand the problem rather than just look for input and output pairs as a typical LLM would.

Llama 2 is the successor to the original Llama LLM, offering improved performance and flexibility. The 13B Chat GPTQ variant is tuned for conversational AI applications optimized for English dialogue.
Some of the models listed above come in multiple spec versions. Generally, higher spec versions will produce better results but require more powerful hardware, while lower spec versions will produce lower quality results but can run on lower-end hardware. If you’re not sure whether your PC can run a model, try the lower spec version first, then move on until you feel the performance drop is no longer acceptable.
The year 2023 saw Microsoft betting heavily on artificial intelligence and its partnership with OpenAI to make Copilot a reality.
Nvidia has just announced the release of an open-source large language model (LLM) that is said to perform on par with leading proprietary models from OpenAI, Anthropic, Meta, and Google.
Foxconn, the company best known for manufacturing iPhones and other Apple hardware products, has just surprised everyone by announcing its first large language model (LLM), called FoxBrain, which is intended to be used to improve manufacturing and supply chain management.
Struggling with Microsoft Teams "Đăng nhập" login error? Discover step-by-step fixes for smooth sign-in. Clear cache, update app, and more – no tech skills needed! Works on Windows, Mac, and web.
Tired of Microsoft Teams Web Error 503 Service Unavailable blocking your meetings? Discover quick, step-by-step fixes to resolve the 503 error fast – no tech skills needed! Clear cache, check status, and get back to collaborating seamlessly.
Stuck with Microsoft Teams Update Error 0x80070002? Discover proven troubleshooting steps to resolve this frustrating issue quickly and get your Teams app updated seamlessly for uninterrupted collaboration.
Tired of Microsoft Teams Error AADSTS50020 blocking your sign-in? Discover step-by-step fixes for user account conflicts, backed by the latest Azure AD solutions. Get back to work fast!
Tired of Microsoft Teams microphone error with no sound? Discover quick, step-by-step fixes for Teams mic not working on Windows, Mac, and more. Restore crystal-clear audio in minutes!
Tired of the Microsoft Teams "Task Management" Error disrupting your workflow? Discover proven fixes like clearing cache, updating Teams, and troubleshooting permissions to get back to seamless collaboration in minutes. Updated with the latest solutions.
Frustrated by the Microsoft Teams 'How to Teams' Help Error? Discover proven, step-by-step solutions to fix it quickly and restore smooth help access. Latest 2026 updates included for seamless teamwork.
Struggling with Microsoft Teams "Room Error" Syncing? This ultimate guide provides step-by-step fixes for Microsoft Teams Room Error Syncing issues, ensuring seamless meetings and quick resolutions. Updated with the latest tips.
Master troubleshooting Microsoft Teams Workflows Power Automate issues with step-by-step fixes for common errors. Get your automations running smoothly – no more frustration! Proven solutions for triggers, permissions, and more.
Struggling with Microsoft Teams "Restart Error" in 2026? Discover proven, step-by-step fixes to resolve the endless restart loop quickly. Clear cache, reset app, and more for seamless collaboration. Get back online fast!
Master how to create breakout rooms in Microsoft Teams before a meeting with this easy guide. Pre-assign participants, boost collaboration, and elevate your virtual sessions effortlessly.
Tired of Microsoft Teams sidebar missing? Get instant fixes for layout errors. Step-by-step guide with latest tips to restore your Teams sidebar fast and frustration-free.
Struggling with Microsoft Teams Download Error on MacBook? Get step-by-step fixes for smooth installation on the latest macOS. Solve download failed issues quickly and easily.
Tired of Microsoft Teams Error J blocking your accounts? Get instant fixes for Error J in Teams with step-by-step guides, cache clears, and account resets. Regain access fast—no tech skills needed!
Work slows down when files sit in too many places, approval steps feel confusing, and people spend more time searching than actually finishing tasks. Strong document management can bring order back without making daily work feel complicated.