Commit graph

404 commits

Author SHA1 Message Date
Simon Willison
9a1374b447
llm embed-multi --prepend option (#746)
* llm embed-multi --prepend option

Closes #745
2025-02-12 15:19:18 -08:00
Simon Willison
f67c21522b
Docs for response.json() and response.usage()
!stable-docs
2025-02-11 08:35:27 -08:00
Simon Willison
41d64a8f12
llm logs --prompts option (#737)
Closes #736
2025-02-02 12:03:01 -08:00
Simon Willison
21df241443 llm-claude-3 is now called llm-anthropic
Refs https://github.com/simonw/llm-claude-3/issues/31

!stable-docs
2025-02-01 22:08:19 -08:00
Simon Willison
f8dcc67455 Release 0.21
Refs #717, #728
2025-01-31 12:35:10 -08:00
Simon Willison
eb0e1e761b o3-mini and reasoning_effort option, refs #728 2025-01-31 12:14:02 -08:00
Simon Willison
656d8fa3c4
--xl/--extract-last flag for prompt and log list commands (#718)
Closes #717
2025-01-24 10:52:46 -08:00
Simon Willison
e449fd4f46
Typo fix
!stable-docs
2025-01-22 22:17:07 -08:00
Simon Willison
3e88628602 uv tool upgrade llm, refs #702
!stable-docs
2025-01-22 21:08:16 -08:00
Simon Willison
bf10f63d3d
Mention gpt-4o-mini-audio-preview too #677
!stable-docs
2025-01-22 21:06:12 -08:00
Simon Willison
eb996baeab Documentation for model.attachment_types, closes #705 2025-01-22 20:46:28 -08:00
Simon Willison
2b9a1bbc50 Fixed broken link 2025-01-22 20:39:01 -08:00
Simon Willison
dc127d2a87 Release 0.20
Refs #654, #676, #677, #681, #688, #690, #700, #702, #709
2025-01-22 20:36:10 -08:00
Simon Willison
57d3baac42 Update embedding model names in docs, refs #654
Also ran Black.
2025-01-22 20:35:17 -08:00
Ryan Patterson
59983740e6
Update directory.md (#666) 2025-01-18 14:52:51 -08:00
abrasumente
e1388b27fe
Add llm-deepseek plugin (#517) 2025-01-11 18:56:34 -08:00
Steven Weaver
2b6b00641c
Update tutorial-model-plugin.md (#685)
pydantic.org -> pydantic.dev
2025-01-11 12:05:05 -08:00
Amjith Ramanujam
e3c104b136
Show the default model when listing all available models. (#688) 2025-01-11 12:04:39 -08:00
Simon Willison
1d75792f9b More uv/uvx tips, closes #702
Refs #690
2025-01-11 10:06:32 -08:00
Ariel Marcus
d964d02e90
Add installation docs with uv (#690) 2025-01-11 09:57:10 -08:00
watany
1c61b5addd
doc(plugin): adding AmazonBedrock (#698) 2025-01-10 16:42:39 -08:00
Arjan Mossel
4f4f9bc07d
Add llm-venice to plugin directory (#699) 2025-01-10 16:41:21 -08:00
Simon Willison
6baf1f7d83 o1
Closes #676
2025-01-10 15:57:06 -08:00
Csaba Henk
88a8cfd9e4
llm logs -x/--extract option (#693)
* llm logs -x/--extract option
* Update docs/help.md for llm logs -x
* Added test for llm logs -x/--extract, refs #693
* llm logs -xr behaves same as llm logs -x
* -x/--extract in llm logging docs

---------

Co-authored-by: Simon Willison <swillison@gmail.com>
2025-01-10 15:53:04 -08:00
Simon Willison
b452effa09 llm models -q/--query option, closes #700 2025-01-09 11:37:33 -08:00
Simon Willison
000e984def --extract support for templates, closes #681 2024-12-19 07:16:48 -08:00
Simon Willison
67d4a99645 llm prompt -x/--extract option, closes #681 2024-12-19 06:40:05 -08:00
Simon Willison
6305b86026 gpt-4o-mini-audio-preview, closes #677 2024-12-17 20:28:57 -08:00
Simon Willison
8898584ba6 New OpenAI audio models, closes #677 2024-12-17 11:14:42 -08:00
Simon Willison
b8e8052229 Release 0.19.1
Refs #667
2024-12-05 13:47:28 -08:00
Simon Willison
e78fea17df Fragment hash on 0.19 release
!stable-docs
2024-12-01 16:09:55 -08:00
Simon Willison
c018104083 Release 0.19
Refs #495, #610, #640, #641, #644, #653
2024-12-01 15:58:27 -08:00
Simon Willison
f9af563df5 response.on_done() mechanism, closes #653 2024-12-01 15:47:23 -08:00
Simon Willison
335b3e635a Release 0.19a2
Refs #640
2024-11-20 20:12:43 -08:00
Simon Willison
c52cfee881 llm.get_models() and llm.get_async_models(), closes #640 2024-11-20 20:09:06 -08:00
Simon Willison
845322e970 Release 0.19a1
Refs #644
2024-11-19 21:28:01 -08:00
Simon Willison
02852fe1a5 Release 0.19a0
Refs #610, #641
2024-11-19 20:23:54 -08:00
Simon Willison
cfb10f4afd
Log input tokens, output tokens and token details (#642)
* Store input_tokens, output_tokens, token_details on Response, closes #610
* llm prompt -u/--usage option
* llm logs -u/--usage option
* Docs on tracking token usage in plugins
* OpenAI default plugin logs usage
2024-11-19 20:21:59 -08:00
Simon Willison
a6d62b7ec9 Release 0.18
Refs #507, #600, #603, #608, #611, #612, #614
2024-11-17 12:31:48 -08:00
Simon Willison
73823012ca Release 0.18a1
Refs #632
2024-11-14 15:10:39 -08:00
Simon Willison
cf172cc70a response.text_or_raise() workaround
Closes https://github.com/simonw/llm/issues/632
2024-11-14 15:08:41 -08:00
Simon Willison
041730d8b2 Release 0.18a0
Refs #507, #599, #600, #603, #608, #611, #612, #613, #614, #615, #616, #621, #622, #623, #626, #629
2024-11-13 17:55:28 -08:00
Simon Willison
ba75c674cb
llm.get_async_model(), llm.AsyncModel base class and OpenAI async models (#613)
- https://github.com/simonw/llm/issues/507#issuecomment-2458639308

* register_model is now async aware

Refs https://github.com/simonw/llm/issues/507#issuecomment-2458658134

* Refactor Chat and AsyncChat to use _Shared base class

Refs https://github.com/simonw/llm/issues/507#issuecomment-2458692338

* fixed function name

* Fix for infinite loop

* Applied Black

* Ran cog

* Applied Black

* Add Response.from_row() classmethod back again

It does not matter that this is a blocking call, since it is a classmethod

* Made mypy happy with llm/models.py

* mypy fixes for openai_models.py

I am unhappy with this, had to duplicate some code.

* First test for AsyncModel

* Still have not quite got this working

* Fix for not loading plugins during tests, refs #626

* audio/wav not audio/wave, refs #603

* Black and mypy and ruff all happy

* Refactor to avoid generics

* Removed obsolete response() method

* Support text = await async_mock_model.prompt("hello")

* Initial docs for llm.get_async_model() and await model.prompt()

Refs #507

* Initial async model plugin creation docs

* duration_ms ANY to pass test

* llm models --async option

Refs https://github.com/simonw/llm/pull/613#issuecomment-2474724406

* Removed obsolete TypeVars

* Expanded register_models() docs for async

* await model.prompt() now returns AsyncResponse

Refs https://github.com/simonw/llm/pull/613#issuecomment-2475157822

---------

Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2024-11-13 17:51:00 -08:00
Hiepler
5a984d0c87
docs: add llm-grok (#629)
Adds`llm-grok` xAI API (https://github.com/Hiepler/llm-grok) to the plugin directory.

!stable-docs
2024-11-13 17:21:04 -08:00
Simon Willison
7520671176 audio/wav not audio/wave, refs #603 2024-11-12 21:43:07 -08:00
Travis Northcutt
c0cb1697bc
Update default model information (#622)
The default model is now 4o-mini; this change updates the usage page of the docs to reflect that
2024-11-12 19:06:16 -08:00
Simon Willison
dff53a9cae Better --help for llm keys get, refs #623 2024-11-11 09:53:24 -08:00
Simon Willison
561784df6e llm keys get command, refs #623 2024-11-11 09:47:13 -08:00
Simon Willison
febbc04fb6
Run cog -r in PRs, use that to update logging.md with new tables (#616)
* Create cog.yml
* Document attachments and prompt_attachments table schemas

Closes #615

---------

Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2024-11-06 06:56:19 -08:00
Simon Willison
98d2c19876 Promote alternative model providers in llm --help 2024-11-06 06:38:53 -08:00
Simon Willison
245e025270 Ran cog, refs #612 2024-11-05 23:45:17 -08:00
Chris Mungall
3b2e5263a3
Allow passing of can_stream in openai_models.py (#600)
* Allow passing of can_stream in openai_models.py

Fixes #599 

* Only set can_stream: false if it is false

Refs https://github.com/simonw/llm/pull/600#issuecomment-2458825866

* Docs for can_stream: false

---------

Co-authored-by: Simon Willison <swillison@gmail.com>
2024-11-05 23:04:13 -08:00
Simon Willison
12df1a3b2a Show attachment types in llm models --options, closes #612 2024-11-05 22:49:26 -08:00
Simon Willison
0cc4072bcd Support attachments without prompts, closes #611 2024-11-05 21:27:18 -08:00
Simon Willison
41cb5c3387 Ran cog, refs #608 2024-11-05 21:13:36 -08:00
Simon Willison
fe1e09706f
llm-lambda-labs
!stable-docs
2024-11-04 10:26:02 -08:00
Simon Willison
a44ba49c21 Release 0.17
Refs #587, #590, #591
2024-10-28 19:36:12 -07:00
Simon Willison
ba1ccb3a4a Release 0.17a0
Refs #587, #590
2024-10-28 15:46:52 -07:00
Simon Willison
1f822d820b Update docs with cog 2024-10-28 15:41:34 -07:00
Simon Willison
f0ed54abf1 Docs for CLI attachments, refs #587 2024-10-28 15:41:34 -07:00
Simon Willison
570a3eccae Python attachment documentation, plus fixed a mimetype detection bug
Refs #587
2024-10-28 15:41:34 -07:00
Simon Willison
1126393ba1 Docs for writing models that accept attachments, refs #587 2024-10-28 15:41:34 -07:00
Simon Willison
7e6031e382
llm-gguf, llm-jq
!stable-docs
2024-10-26 22:44:06 -07:00
Simon Willison
d654c95212 Release notes for 0.16 2024-09-12 16:20:12 -07:00
Simon Willison
bfcfd2c91b
o1-preview and o1-mini, refs #570 (#573) 2024-09-12 16:08:04 -07:00
Kian-Meng Ang
50520c7c1c
Fix typos (#567)
Found via `codespell -H -L wit,thre`

!stable-docs
2024-09-08 08:44:43 -07:00
Simon Willison
7d6ece2a31 Fix for broken markdown on openai-models page
Refs #558 !stable-docs
2024-08-25 18:03:46 -07:00
Simon Willison
6deed8f976 get_model() improvement, get_default_model() / set_default_wodel() now documented
Refs #553
2024-08-18 17:37:31 -07:00
Simon Willison
d075336c69 Release 0.15
Refs #515, #525, #536, #537
2024-07-18 12:31:14 -07:00
Simon Willison
562fefb374 Use 3-small in docs instead of ada-002
Spotted while working on #537
2024-07-18 12:23:49 -07:00
Simon Willison
fcba89d73b Update docs to reflect new gpt-4o-mini default, refs #536 2024-07-18 12:16:03 -07:00
Simon Donohue
50454c1957
Update outdated reference to gpt-4-turbo (#525)
Looks like this alias was overlooked in 8171c9a. This commit makes it
match with the usage of gpt-4o in the associated example.
2024-07-18 12:10:40 -07:00
Simon Willison
2881576dd0 Re-ran cog, refs #536 2024-07-18 12:00:35 -07:00
Simon Willison
96db13f537
Link to new video
!stable-docs
2024-06-17 10:18:24 -07:00
Simon Willison
68df9721de
github repo static badge
!stable-docs
2024-05-13 18:41:07 -07:00
Simon Willison
45245413bd
GitHub stars badge
!stable-docs
2024-05-13 15:09:56 -07:00
Simon Willison
9a3236db61 gpt-4-turbo model ID, closes #493 2024-05-13 13:37:23 -07:00
Simon Willison
ab1cc4fd1f Release 0.14
Refs #404, #431, #470, #490, #491
2024-05-13 13:26:48 -07:00
Fabian Labat
6cdc29c8d6
Update directory.md (#486)
* Update directory.md

Added support for Bedrock Llama 3
2024-05-13 13:01:33 -07:00
Simon Willison
3cc588f247 List llm-llamafile in plugins directory, closes #470 2024-05-13 12:55:22 -07:00
Simon Willison
8171c9a6bf Update help for GPT-4o, closes #490 2024-05-13 12:53:31 -07:00
Simon Willison
73bbbec372 gpt-4o model, refs #490 2024-05-13 12:49:45 -07:00
Simon Willison
04915e95f8
llm-groq
!stable-docs
2024-04-21 20:33:23 -07:00
Simon Willison
2a9b6113f5
llm-perplexity
Refs https://github.com/hex/llm-perplexity/issues/2

!stable-docs
2024-04-21 16:18:37 -07:00
Simon Willison
99a2836638
llm-fireworks
Refs https://github.com/simonw/llm-fireworks/issues/1

!stable-docs
2024-04-18 17:20:09 -07:00
Simon Willison
9ad9ac68dc
llm-reka in plugin directory
!stable-docs
2024-04-17 19:38:41 -07:00
Simon Willison
12e027d3e4
llm-command-r
!stable-docs

Refs https://github.com/simonw/llm-command-r/issues/1
2024-04-04 07:41:03 -07:00
Simon Willison
008efae86a
llm-cmd
!stable-docs

Refs https://github.com/simonw/llm-cmd/issues/1
2024-03-26 08:58:48 -07:00
Simon Willison
fb63c92cd2 llm logs -r/--response option, closes #431 2024-03-04 13:29:07 -08:00
Simon Willison
5852aa1574
llm-claude-3
!stable-docs
2024-03-04 10:48:57 -08:00
Simon Willison
de6af1ca2f
llm-embed-onnx in plugin directory
!stable-docs

Refs https://github.com/simonw/llm-embed-onnx/issues/1
2024-01-28 14:27:13 -08:00
Simon Willison
5293b9dd72
Added llm-together
!stable-docs
2024-01-27 10:40:40 -08:00
Simon Willison
7e89b88565 Fixed missing link in changelog for 0.4
!stable-docs
2024-01-26 21:54:37 -08:00
Simon Willison
8021e12aaa
Windows readline fix, plus run CI against macOS and Windows
* Run CI on Windows and macOS as well as Ubuntu, refs #407
* Use pyreadline3 on win32
* Back to fail-fast since we have a bigger matrix now
* Mark some tests as xfail on windows
2024-01-26 16:24:58 -08:00
Simon Willison
e32d9b35e3
Link to annotated release notes
!stable-docs
2024-01-26 15:11:56 -08:00
Simon Willison
8e0aff6090 Release 0.13
Refs #325, #351, #373, #376, #394, #396, #397, #404

Closes #402
2024-01-26 14:32:23 -08:00
Fabian Labat
81225ab94b
llm-bedrock-meta plugin
!stable-docs
2024-01-26 14:13:20 -08:00
Simon Willison
2b92756046 Homebrew PyTorch warning, closes #397 2024-01-26 13:59:27 -08:00
Simon Willison
46763326ed Homebrew installation, refs #315, #397
This reverts commit abcb457b20.
2024-01-26 13:50:21 -08:00
Simon Willison
4f7489876d Syntax highlighting for docs/contributing.md 2024-01-26 12:05:58 -08:00