Tag: llm
All the articles with the tag "llm".
-
LLM Backends: vLLM vs llama.cpp vs Ollama
vLLM, llama.cpp, and Ollama all run local LLMs — compare throughput, memory use, GPU support, and which fits your hardware.
-
Key Parameters of Large Language Models
Temperature, top-p, top-k, context length — LLM inference parameters explained so you stop guessing why the model gives weird output.
-
Prompt Engineering for Generative AI 101
Write prompts that get useful results — role prompting, few-shot examples, chain-of-thought, and the patterns that work across any LLM.
-
Large Language Model Formats and Quantization
GGUF, GGML, AWQ, GPTQ — LLM file formats and quantization levels explained: trade-offs between model quality, size, and inference speed.
-
Exploring the Diverse World of LLM Models
LLaMA, Mistral, Falcon, GPT — the LLM landscape is crowded. Compare model families, sizes, licensing, and what each is actually good for.
-
Ollama: Powerful Language Models on Your Own Machine
Ollama makes running local LLMs dead simple — pull a model, start the server, and get a private ChatGPT running on your own hardware.
-
Unleash the Power of LLMs with LocalAI
LocalAI is a self-hosted OpenAI-compatible API — run any GGUF model and connect existing tools without changing a line of client code.
-
Machine Learning models (AI)
Supervised, unsupervised, reinforcement learning — the ML model landscape explained without drowning in math or hype.