Skip to main content

3 posts tagged with "LLM"

View All Tags

Configuring Groq with Llama Models

· 3 min read
Anand Raja
Senior Software Engineer

Groq offers incredibly fast LLM inference through its API service, including access to Llama models. Here's a complete guide to setting up and using Groq with Llama models at no cost.

What is Groq?

Groq is an AI inference engine that provides remarkably fast LLM responses using specialized hardware accelerators (LPUs - Language Processing Units). Their free tier allows access to several models, including Meta's Llama family.

Ollama: Usage Notes & Quick Reference

· 10 min read
Anand Raja
Senior Software Engineer

Ollama is an open-source tool that is designed to simplify the process of running large language models locally, meaning on your own hardware. The idea here is very simple. As you know, if you want to run large language models or use a model, most likely you will have to rely on paid services like OpenAI, ChatGPT, and others.

With Ollama, you don't have to pay for anything—it's free, and that's the beauty. Ollama sits at the center and allows developers to pick different large language models depending on the situation and their needs. At its core, Ollama uses a command-line interface (CLI) to manage backend tasks like installation and execution of different models, all of which run locally. Ollama abstracts away the technical complexities involved in setting up these models, making advanced language processing accessible to a broader audience, including developers, researchers, and hobbyists. In a nutshell, Ollama provides a straightforward way to download, run, and interact with various models or LLMs without relying on cloud-based services or dealing with complex setup procedures.

Open LLMs: Concepts, Parameters, Ecosystem, and Usage

· 17 min read
Anand Raja
Senior Software Engineer

What is an Open LLM?

An Open LLM (Large Language Model) is a language model whose architecture, weights, and often training code are openly available for anyone to use, modify, and run. Unlike proprietary LLMs (like OpenAI’s GPT-4 or Google Gemini), open LLMs can be run locally, fine-tuned, and integrated into custom workflows without vendor lock-in or sending your data to third-party servers.