mirror of
https://gitea.com/mcereda/oam.git
synced 2026-03-05 00:14:25 +00:00
chore(kb/ai): review and expand notes
This commit is contained in:
@@ -177,6 +177,7 @@ See [An AI Agent Published a Hit Piece on Me] by Scott Shambaugh.
|
||||
- [The 2026 Guide to Coding CLI Tools: 15 AI Agents Compared]
|
||||
- [Evaluating AGENTS.md: Are Repository-Level Context Files Helpful for Coding Agents?]
|
||||
- [SkillsBench: Benchmarking How Well Agent Skills Work Across Diverse Tasks]
|
||||
- [AI mistakes you're probably making]
|
||||
|
||||
### Sources
|
||||
|
||||
@@ -208,6 +209,7 @@ See [An AI Agent Published a Hit Piece on Me] by Scott Shambaugh.
|
||||
[Agent Skills]: https://agentskills.io/
|
||||
[Agentic ProbLLMs - The Month of AI Bugs]: https://monthofaibugs.com/
|
||||
[AI Doesn't Reduce Work — It Intensifies It]: https://hbr.org/2026/02/ai-doesnt-reduce-work-it-intensifies-it
|
||||
[AI mistakes you're probably making]: https://www.youtube.com/watch?v=Jcuig8vhmx4
|
||||
[An AI Agent Published a Hit Piece on Me]: https://theshamblog.com/an-ai-agent-published-a-hit-piece-on-me/
|
||||
[ASCII Smuggler Tool: Crafting Invisible Text and Decoding Hidden Codes]: https://embracethered.com/blog/posts/2024/hiding-and-finding-text-with-unicode-tags/
|
||||
[Evaluating AGENTS.md: Are Repository-Level Context Files Helpful for Coding Agents?]: https://arxiv.org/abs/2602.11988
|
||||
|
||||
@@ -1,53 +1,62 @@
|
||||
# Claude
|
||||
|
||||
> TODO
|
||||
|
||||
AI platform built by Anthropic.
|
||||
|
||||
<!-- Remove this line to uncomment if used
|
||||
## Table of contents <!-- omit in toc -->
|
||||
Family of [LLMs][large language models] developed by Anthropic.
|
||||
|
||||
1. [TL;DR](#tldr)
|
||||
1. [Models' code of conduct](#models-code-of-conduct)
|
||||
1. [Further readings](#further-readings)
|
||||
1. [Sources](#sources)
|
||||
|
||||
## TL;DR
|
||||
|
||||
<!-- Uncomment if used
|
||||
<details>
|
||||
<summary>Setup</summary>
|
||||
As of 2026-03-02, all models support text and image input, text output, multilingual capabilities, and vision.
|
||||
|
||||
```sh
|
||||
```
|
||||
Prefer **Opus** for the most _demanding_ tasks or when in need for deep reasoning, e.g. large-scale code refactoring,
|
||||
complex architectural decisions, multi-step research and analysis, or advanced agentic workflows.<br/>
|
||||
It is built to excel at coding and complex problem-solving, and to tackle sustained performance on long-running tasks
|
||||
that span multiple of steps over several hours.<br/>
|
||||
It is also the **most** expensive of Anthropic's models.
|
||||
|
||||
</details>
|
||||
-->
|
||||
Prefer **Haiku** for near-real-time responses and/or high-volume, lower-complexity tasks, e.g. classifying feedback,
|
||||
summarizing support tickets, lightweight retrieval-augmented answers, and in-product micro-interactions.<br/>
|
||||
It is the **least** expensive of Anthropic's models.
|
||||
|
||||
<!-- Uncomment if used
|
||||
<details>
|
||||
<summary>Usage</summary>
|
||||
Prefer **Sonnet** when wanting to balance speed and reasoning capabilities, handling everyday coding, writing, analysis,
|
||||
summarization, and document work.<br/>
|
||||
It is usually fast and reliable enough for everyday work, and can switch to deeper thinking when tasks get harder.
|
||||
|
||||
```sh
|
||||
```
|
||||
When in doubt, start with Sonnet, then consider changing model should Sonnet fly through task (then maybe Haiku is
|
||||
enough) or have troubles with them (escalating to Opus).
|
||||
|
||||
</details>
|
||||
-->
|
||||
Anthropic is pushing its models to interiorize a sort of [code of conduct][models' code of conduct].
|
||||
|
||||
<!-- Uncomment if used
|
||||
<details>
|
||||
<summary>Real world use cases</summary>
|
||||
## Models' code of conduct
|
||||
|
||||
```sh
|
||||
```
|
||||
Anthropic trains its models with a code of conduct of sorts during training to shape its values and judgement.<br/>
|
||||
The goal is for Claude to internalize good principles deeply enough to generalize to new situations. Some behaviors
|
||||
should be absolute hard limits (e.g., never help with bioweapons), others should be adjustable defaults that operators
|
||||
and users can modify _within bounds_.
|
||||
|
||||
</details>
|
||||
-->
|
||||
Refer to [Claude's Constitution].
|
||||
|
||||
Claude models are expected to:
|
||||
|
||||
1. Be **_broadly_ safe** by supporting human oversight of AI during the early period of development.
|
||||
1. Be **_broadly_ ethical** by being honest, acting according to good values and intentions, and avoiding actions that
|
||||
are inappropriate, dangerous, or harmful.
|
||||
1. **Comply with Anthropic's guidelines** where relevant.
|
||||
1. Be **_genuinely_ helpful** by providing real value to users
|
||||
|
||||
In cases of apparent conflict, models should _generally_ prioritize these properties **in the order in which they're
|
||||
listed**.
|
||||
|
||||
## Further readings
|
||||
|
||||
- [Website]
|
||||
- [Blog]
|
||||
- [Pricing]
|
||||
- [Large Language Models]
|
||||
- [Claude's Constitution]
|
||||
- [Gemini]
|
||||
|
||||
### Sources
|
||||
@@ -60,6 +69,8 @@ AI platform built by Anthropic.
|
||||
-->
|
||||
|
||||
<!-- In-article sections -->
|
||||
[Models' code of conduct]: #models-code-of-conduct
|
||||
|
||||
<!-- Knowledge base -->
|
||||
[Gemini]: ../gemini/README.md
|
||||
[Large Language Models]: ../lms.md#large-language-models
|
||||
@@ -67,7 +78,9 @@ AI platform built by Anthropic.
|
||||
<!-- Files -->
|
||||
<!-- Upstream -->
|
||||
[Blog]: https://claude.com/blog
|
||||
[Claude's Constitution]: https://www.anthropic.com/constitution
|
||||
[Developer documentation]: https://platform.claude.com/docs/en/home
|
||||
[Pricing]: https://claude.com/pricing
|
||||
[Website]: https://claude.com/product/overview
|
||||
|
||||
<!-- Others -->
|
||||
|
||||
Reference in New Issue
Block a user