From 6754a3261a6c7cda9a0f615958e8033eba16d330 Mon Sep 17 00:00:00 2001 From: Michele Cereda Date: Fri, 6 Feb 2026 00:11:05 +0100 Subject: [PATCH] feat(kb/ai): add llm article --- knowledge base/ai/README.md | 9 +- knowledge base/ai/large language model.md | 104 ++++++++++++++++++ ...cal llm for coding assistance in vscode.md | 2 + knowledge base/ai/vllm.md | 67 +++++++++++ knowledge base/jargon.md | 3 +- 5 files changed, 178 insertions(+), 7 deletions(-) create mode 100644 knowledge base/ai/large language model.md create mode 100644 knowledge base/ai/vllm.md diff --git a/knowledge base/ai/README.md b/knowledge base/ai/README.md index 9595345..b0fcb16 100644 --- a/knowledge base/ai/README.md +++ b/knowledge base/ai/README.md @@ -17,8 +17,8 @@ TODO ## Further readings +- [Large Language Model] (LLM) - [Useful AI]: tools, courses, and more, curated and reviewed by experts. -- LLMs: [ChatGPT], [Claude], [Copilot], [Duck AI], [Gemini] ### Sources @@ -29,12 +29,9 @@ TODO +[Large Language Model]: large%20language%20model.md + -[ChatGPT]: https://chatgpt.com/ -[Claude]: https://claude.ai/ -[Copilot]: https://copilot.microsoft.com/ -[Duck AI]: https://duck.ai/c -[Gemini]: https://gemini.google.com/ [Useful AI]: https://usefulai.com/ diff --git a/knowledge base/ai/large language model.md b/knowledge base/ai/large language model.md new file mode 100644 index 0000000..6827744 --- /dev/null +++ b/knowledge base/ai/large language model.md @@ -0,0 +1,104 @@ +# Large language model + +_Language models_ are computational model that can predict sequences in natural language.
+Useful for speech recognition, machine translation, natural language generation, optical character recognition, route +optimization, handwriting recognition, grammar induction, information retrieval, and other tasks. + +_Large_ language models are predominantly based on transformers trained on large datasets, frequently including texts +scraped from the Internet.
+They have superseded recurrent neural network-based models. + + + +1. [TL;DR](#tldr) +1. [Run LLMs Locally](#run-llms-locally) +1. [Further readings](#further-readings) + 1. [Sources](#sources) + +## TL;DR + +| FIXME | Creator | +| --------- | ---------- | +| [ChatGPT] | OpenAI | +| [Claude] | Anthropic | +| [Copilot] | Microsoft | +| [Duck AI] | DuckDuckGo | +| [Gemini] | Google | +| [Grok] | X | +| [Llama] | Meta | +| [Mistral] | Mistral AI | + + + + + + + +## Run LLMs Locally + +Use one of the following: + +- [Ollama] +- [LMStudio] +- [vLLM] +- [Jan] +- [llama.cpp] +- [Llamafile] + +## Further readings + +### Sources + +- [Run LLMs Locally: 6 Simple Methods] + + + + + +[LMStudio]: lmstudio.md +[Ollama]: ollama.md +[vLLM]: vllm.md + + + + +[ChatGPT]: https://chatgpt.com/ +[Claude]: https://claude.ai/ +[Copilot]: https://copilot.microsoft.com/ +[Duck AI]: https://duck.ai/ +[Gemini]: https://gemini.google.com/ +[Grok]: https://grok.com/ +[Jan]: https://www.jan.ai/ +[llama.cpp]: https://github.com/ggml-org/llama.cpp +[Llama]: https://www.llama.com/ +[Llamafile]: https://github.com/mozilla-ai/llamafile +[Mistral]: https://mistral.ai/ +[Run LLMs Locally: 6 Simple Methods]: https://www.datacamp.com/tutorial/run-llms-locally-tutorial diff --git a/knowledge base/ai/use a local llm for coding assistance in vscode.md b/knowledge base/ai/use a local llm for coding assistance in vscode.md index 66e7491..8dd740f 100644 --- a/knowledge base/ai/use a local llm for coding assistance in vscode.md +++ b/knowledge base/ai/use a local llm for coding assistance in vscode.md @@ -33,6 +33,7 @@ ## Further readings +- [Large Language Model] (LLM) - [Ollama] - [Continue VSCode extension] @@ -47,6 +48,7 @@ +[Large Language Model]: large%20language%20model.md [Ollama]: ollama.md diff --git a/knowledge base/ai/vllm.md b/knowledge base/ai/vllm.md new file mode 100644 index 0000000..2c4cb53 --- /dev/null +++ b/knowledge base/ai/vllm.md @@ -0,0 +1,67 @@ +# Title + +Open source library for LLM inference and serving. + + + +1. [TL;DR](#tldr) +1. [Further readings](#further-readings) + 1. [Sources](#sources) + +## TL;DR + +
+ Setup + +```sh +pip install 'vllm' +``` + +
+ +
+ Usage + +```sh +vllm serve 'meta-llama/Llama-2-7b-hf' --port '8000' --gpu-memory-utilization '0.9' +vllm serve 'meta-llama/Llama-2-70b-hf' --tensor-parallel-size '2' --port '8000' +``` + +
+ + + +## Further readings + +- [Website] +- [Codebase] +- [Blog] + +### Sources + +- [Documentation] + + + + + + + +[Blog]: https://blog.vllm.ai/ +[Codebase]: https://github.com/vllm-project/vllm +[Documentation]: https://docs.vllm.ai/en/ +[Website]: https://vllm.ai/ + + diff --git a/knowledge base/jargon.md b/knowledge base/jargon.md index 55dfb92..3032470 100644 --- a/knowledge base/jargon.md +++ b/knowledge base/jargon.md @@ -90,7 +90,7 @@ | LAN | Local Area Network | | | LED | Light Emitting Diode | | | LIFO | Last In First Out | | -| LLM | Large Language Model | | +| LLM | [Large Language Model] | | | M2COTS | Mass Market COTS | Widely available COTS products | | MR | Merge Request | Prevalently used in GitLab | | NACL | Network ACL | | @@ -182,6 +182,7 @@ [fish]: fish.md [iac]: iac.md [kubernetes]: kubernetes/README.md +[Large Language Model]: ai/large%20language%20model.md [lora]: lora.md [siem]: siem.md [snowflake]: snowflake/README.md