mirror of
https://gitea.com/mcereda/oam.git
synced 2026-03-01 14:34:24 +00:00
chore(kb/ai): review and expand notes
This commit is contained in:
@@ -70,6 +70,10 @@ Best practices:
|
||||
- Consider limiting agent execution to containers or otherwise isolated environments, with only (limited) access to
|
||||
what they _absolutely_ need.
|
||||
- Prefer **requiring** consent by agents when running them.
|
||||
- Prefer including **only minimal requirements** in context files (AGENTS.md).<br/>
|
||||
Too much context ends up hurting the conversation. Including a lot of "don't do this or that" mostly poisons the
|
||||
context instead of helping.
|
||||
If an information is in the codebase, it probably does not need to be in the context file.
|
||||
|
||||
## Skills
|
||||
|
||||
@@ -171,6 +175,8 @@ See [An AI Agent Published a Hit Piece on Me] by Scott Shambaugh.
|
||||
- [Token Anxiety] by Nikunj Kothari
|
||||
- [AI Doesn't Reduce Work — It Intensifies It] by Aruna Ranganathan and Xingqi Maggie Ye
|
||||
- [The 2026 Guide to Coding CLI Tools: 15 AI Agents Compared]
|
||||
- [Evaluating AGENTS.md: Are Repository-Level Context Files Helpful for Coding Agents?]
|
||||
- [SkillsBench: Benchmarking How Well Agent Skills Work Across Diverse Tasks]
|
||||
|
||||
### Sources
|
||||
|
||||
@@ -204,6 +210,7 @@ See [An AI Agent Published a Hit Piece on Me] by Scott Shambaugh.
|
||||
[AI Doesn't Reduce Work — It Intensifies It]: https://hbr.org/2026/02/ai-doesnt-reduce-work-it-intensifies-it
|
||||
[An AI Agent Published a Hit Piece on Me]: https://theshamblog.com/an-ai-agent-published-a-hit-piece-on-me/
|
||||
[ASCII Smuggler Tool: Crafting Invisible Text and Decoding Hidden Codes]: https://embracethered.com/blog/posts/2024/hiding-and-finding-text-with-unicode-tags/
|
||||
[Evaluating AGENTS.md: Are Repository-Level Context Files Helpful for Coding Agents?]: https://arxiv.org/abs/2602.11988
|
||||
[Forget the Hype: Agents are Loops]: https://dev.to/cloudx/forget-the-hype-agents-are-loops-1n3i
|
||||
[How a Single Email Turned My ClawdBot Into a Data Leak]: https://medium.com/@peltomakiw/how-a-single-email-turned-my-clawdbot-into-a-data-leak-1058792e783a
|
||||
[moltbot security situation is insane]: https://www.youtube.com/watch?v=kSno1-xOjwI
|
||||
@@ -212,12 +219,13 @@ See [An AI Agent Published a Hit Piece on Me] by Scott Shambaugh.
|
||||
[OpenClaw: Who are you?]: https://www.youtube.com/watch?v=hoeEclqW8Gs
|
||||
[openclaw/openclaw]: https://github.com/openclaw/openclaw
|
||||
[Remote Labor Index: Measuring AI Automation of Remote Work]: https://arxiv.org/abs/2510.26787
|
||||
[SkillsBench: Benchmarking How Well Agent Skills Work Across Diverse Tasks]: https://arxiv.org/abs/2602.12670
|
||||
[Stealing everything you've ever typed or viewed on your own Windows PC is now possible with two lines of code — inside the Copilot+ Recall disaster.]: https://doublepulsar.com/recall-stealing-everything-youve-ever-typed-or-viewed-on-your-own-windows-pc-is-now-possible-da3e12e9465e
|
||||
[Superpowers: How I'm using coding agents in October 2025]: https://blog.fsck.com/2025/10/09/superpowers/
|
||||
[The 2026 Guide to AI Agents]: https://www.ibm.com/think/ai-agents
|
||||
[The 2026 Guide to Coding CLI Tools: 15 AI Agents Compared]: https://www.tembo.io/blog/coding-cli-tools-comparison
|
||||
[The Agentic Loop, Explained: What Every PM Should Know About How AI Agents Actually Work]: https://www.ikangai.com/the-agentic-loop-explained-what-every-pm-should-know-about-how-ai-agents-actually-work/
|
||||
[Token Anxiety]: https://writing.nikunjk.com/p/token-anxiety
|
||||
[TotalRecall]: https://github.com/xaitax/TotalRecall
|
||||
[Trust No AI: Prompt Injection Along The CIA Security Triad]: https://arxiv.org/pdf/2412.06090
|
||||
[xAI engineer fired for leaking secret "Human Emulator" project]: https://www.youtube.com/watch?v=0hDMSS1p-UY
|
||||
[The 2026 Guide to Coding CLI Tools: 15 AI Agents Compared]: https://www.tembo.io/blog/coding-cli-tools-comparison
|
||||
|
||||
@@ -9,11 +9,12 @@ the previous ones.
|
||||
1. [Speculative decoding](#speculative-decoding)
|
||||
1. [Reasoning](#reasoning)
|
||||
1. [Prompting](#prompting)
|
||||
1. [Context window](#context-window)
|
||||
1. [Function calling](#function-calling)
|
||||
1. [Concerns](#concerns)
|
||||
1. [Run LLMs Locally](#run-llms-locally)
|
||||
1. [Further readings](#further-readings)
|
||||
1. [Sources](#sources)
|
||||
1. [Sources](#sources)
|
||||
|
||||
## TL;DR
|
||||
|
||||
@@ -34,7 +35,9 @@ _Context_ is helpful information before or after a target token.<br/>
|
||||
It can help a language model make better predictions, like determining whether "orange" refers to a citrus fruit or a
|
||||
color.
|
||||
|
||||
_Large_ LMs are language models trained on massive datasets, and encoding their acquired knowledge into up to trillions
|
||||
_Context Window_ is the amount of tokens that a model can pay attention to at any one time.
|
||||
|
||||
_Large LMs_ are language models trained on massive datasets, and encoding their acquired knowledge into up to trillions
|
||||
of parameters.
|
||||
|
||||
_Parameters_ are internal weights and values that an LLM learns during training.<br/>
|
||||
@@ -44,6 +47,8 @@ The more parameters a model has, the better it typically is to understand and ge
|
||||
An increased parameter count, on the other hand, demands more computational resources for training and inference, and
|
||||
make models more prone to overfitting, slower to respond, and harder to deploy efficiently.
|
||||
|
||||
_System prompts_ are predefined text included at the start of conversations to establish ground rules for them.
|
||||
|
||||
| Provider | Creator |
|
||||
| --------- | ---------- |
|
||||
| [ChatGPT] | OpenAI |
|
||||
@@ -203,6 +208,62 @@ Users can confirm the plan **before** the model executes it.
|
||||
_Good_ prompting is about designing predictable interactions with a model.<br/>
|
||||
In the context of LLM agent development, it is no different from interface design.
|
||||
|
||||
Model providers and creators usually bake behavioural rules into the model through training, rather than just using
|
||||
instructions.<br/>
|
||||
These aren't _rules that a model follows_ so much as _values_ that are deeply embedded in how it thinks and
|
||||
responds.<br/>
|
||||
They act as the limits that **no** prompt or user request can override.
|
||||
|
||||
_System prompts_ are sets of guidelines, boundaries, and contextual information that establishes ground rules for the
|
||||
conversations to come.<br/>
|
||||
They guide the model to operate within specific parameters and generate responses that are coherent,
|
||||
relevant, and aligned with the desired outcome.
|
||||
|
||||
<details style='padding: 0 0 1rem 1rem'>
|
||||
<summary>Example</summary>
|
||||
|
||||
```md
|
||||
You are an enthusiastic biology teacher named Leonardo.
|
||||
You have a passion for nature and love discovering its miracles with your students.
|
||||
Your communication style is friendly and informative.
|
||||
|
||||
If a user asks about topics outside your expertise, such as medical advice or legal matters, inform them that you are
|
||||
not qualified to provide relevant responses. Suggest that they should consult with the appropriate professionals.
|
||||
```
|
||||
|
||||
</details>
|
||||
|
||||
System prompts are designed to take precedence over user instructions.<br/>
|
||||
If a system prompt says to "never discuss competitor products", users asking about competitors should still be denied.
|
||||
However, users can sometimes work around poorly written system prompts through clever phrasing or _jailbreaking_
|
||||
attempts.
|
||||
|
||||
Also see [Mastering System Prompts for LLMs] for suggestions.
|
||||
|
||||
## Context window
|
||||
|
||||
Amount of text, **in tokens**, that a model can _remember_ at any one time.<br/>
|
||||
Determines how long of a conversation it can carry out without forgetting details from earlier in the exchange, and the
|
||||
maximum size of documents or code samples that it can process at once.
|
||||
|
||||
Generally, increasing an LLM's context size increases accuracy and coherent responses, lowers hallucinations, allows for
|
||||
longer conversations, and improves analysis of longer sequences of data.<br/>
|
||||
Larger windows require more computational resources (especially vRAM), and can overwhelm the model pushing them to take
|
||||
cognitive shortcuts. This potentially increases a model's vulnerability to manipulation.
|
||||
|
||||
When sending messages to a model, one is really sending **the whole context** up to the current point in the
|
||||
conversation (the message history, documents, and all).
|
||||
|
||||
Models perform best when relevant information is toward the **beginning** or **end** of the input context.<br/>
|
||||
Performance degrades when the model must carefully consider the information **in the middle** of long contexts.
|
||||
|
||||
When a prompt, conversation, document or code base exceeds a model's context window, the context must be _compacted_ to
|
||||
continue.<br/>
|
||||
The compacting action summarizes the conversation, then clears the current context and reloads that summary. The model
|
||||
retains _an idea_ of the progress, and now can continue.
|
||||
|
||||
Whatever one puts in the context, that is more likely to either happen or to steer the conversation towards it.
|
||||
|
||||
## Function calling
|
||||
|
||||
Refer [Function calling in LLMs].
|
||||
@@ -285,6 +346,7 @@ Refer:
|
||||
- [Introduction to Large Language Models]
|
||||
- GeeksForGeeks' [What are LLM parameters?][geeksforgeeks / what are llm parameters?]
|
||||
- IBM's [What are LLM parameters?][ibm / what are llm parameters?]
|
||||
and [What is a context window?][ibm / what is a context window?]
|
||||
- [This is not the AI we were promised], presentation by Michael John Wooldridge at the Royal Society
|
||||
|
||||
<!--
|
||||
@@ -318,6 +380,7 @@ Refer:
|
||||
[GeeksForGeeks / What are LLM parameters?]: https://www.geeksforgeeks.org/artificial-intelligence/what-are-llm-parameters/
|
||||
[Grok]: https://grok.com/
|
||||
[IBM / What are LLM parameters?]: https://www.ibm.com/think/topics/llm-parameters
|
||||
[IBM / What is a context window?]: https://www.ibm.com/think/topics/context-window
|
||||
[Introduction to Large Language Models]: https://developers.google.com/machine-learning/crash-course/llm
|
||||
[Jan]: https://www.jan.ai/
|
||||
[Llama]: https://www.llama.com/
|
||||
@@ -325,6 +388,7 @@ Refer:
|
||||
[LLM skills every AI engineer must know]: https://fiodar.substack.com/p/llm-skills-every-ai-engineer-must-know
|
||||
[Local LLM Hosting: Complete 2026 Guide - Ollama, vLLM, LocalAI, Jan, LM Studio & More]: https://www.glukhov.org/post/2025/11/hosting-llms-ollama-localai-jan-lmstudio-vllm-comparison/
|
||||
[Looking back at speculative decoding]: https://research.google/blog/looking-back-at-speculative-decoding/
|
||||
[Mastering System Prompts for LLMs]: https://dev.to/simplr_sh/mastering-system-prompts-for-llms-2d1d
|
||||
[Mistral]: https://mistral.ai/
|
||||
[OpenClaw: Who are you?]: https://www.youtube.com/watch?v=hoeEclqW8Gs
|
||||
[Optimizing LLMs for Performance and Accuracy with Post-Training Quantization]: https://developer.nvidia.com/blog/optimizing-llms-for-performance-and-accuracy-with-post-training-quantization/
|
||||
|
||||
Reference in New Issue
Block a user