Commit graph

76 commits

Author SHA1 Message Date
Simon Willison
e48a5b9f11 Rename "echo" to "simple-echo", refs #1061 2025-05-20 20:55:54 -07:00
Sukhbinder Singh
8c7d33ee52
Fixes --continue bug and adds --database argument to llm chat
* Fix database bug in continue conversation and adds --database to llm chat
* Move the --database to proper place and update help.

Closes #933
2025-05-04 15:57:13 -07:00
Simon Willison
e78e1fceb2 LLM_MODEL and LLM_EMBEDDING_MODEL environment vars, closes #932 2025-04-19 20:41:24 -07:00
Steve Morin
4894b13414
llm models --options now shows keys and environment variables
Closes #903
2025-04-08 20:15:56 -07:00
Simon Willison
f740a5cbbd
Fragments (#859)
* WIP fragments: schema plus reading but not yet writing, refs #617
* Unique index on fragments.alias, refs #617
* Fragments are now persisted, added basic CLI commands
* Fragment aliases work now, refs #617
* Improved help for -f/--fragment
* Support fragment hash as well
* Documentation for fragments
* Better non-JSON display of llm fragments list
* llm fragments -q search option
* _truncate_string is now truncate_string
* Use condense_json to avoid duplicate data in JSON in DB, refs #617
* Follow up to 3 redirects for fragments
* Python API docs for fragments= and system_fragments=
* Fragment aliases cannot contain a : - this is to ensure we can add custom fragment loaders later on, refs https://github.com/simonw/llm/pull/859#issuecomment-2761534692
* Use template fragments when running prompts
* llm fragments show command plus llm fragments group tests
* Tests for fragments family of commands
* Test for --save with fragments
* Add fragments tables to docs/logging.md
* Slightly better llm fragments --help
* Handle fragments in past conversations correctly
* Hint at llm prompt --help in llm --help, closes #868
* llm logs -f filter plus show fragments in llm logs --json
* Include prompt and system fragments in llm logs -s
* llm logs markdown fragment output and tests, refs #617
2025-04-05 17:22:37 -07:00
Simon Willison
5b2c611c82 llm prompt -d/--database option, closes #858 2025-03-28 00:20:31 -07:00
Simon Willison
3541415db4 llm prompt -q X -q Y option, closes #841 2025-03-21 15:17:16 -07:00
Simon Willison
63e5dff774 Update test, regs #834 2025-03-19 20:47:51 -07:00
Simon Willison
1d552aeacc llm models -m option, closes #825 2025-03-10 14:18:50 -07:00
Simon Willison
9a38021218 llm schemas dsl command, closes #793
Refs #790
2025-02-27 10:46:56 -08:00
Simon Willison
7e819c2ffa Implemented --schema-multi, closes #791 2025-02-27 10:12:21 -08:00
Simon Willison
321636e791 New schema DSL, closes #790
Plus made a start on schemas.md refs #788
2025-02-27 09:48:44 -08:00
Simon Willison
f35ac31c21
llm logs --schema, --data, --data-array and --data-key options (#785)
* llm logs --schema option, refs #782
* --data and --data-array and --data-key options, refs #782
* Tests for llm logs --schema options, refs #785
* Also implemented --schema ID lookup, refs #780
* Using --data-key implies --data
* Docs for llm logs --schema and --data etc
2025-02-26 21:51:08 -08:00
Simon Willison
02999e398d Refactor tests to new test_llm_logs.py module
Refs #785
2025-02-26 20:23:55 -08:00
Simon Willison
3d4871f163 Improved log Markdown, closes #783 2025-02-26 19:25:25 -08:00
Simon Willison
62c90dd472
llm prompt --schema X option and model.prompt(..., schema=) parameter (#777)
Refs #776

* Implemented new llm prompt --schema and model.prompt(schema=)
* Log schema to responses.schema_id and schemas table
* Include schema in llm logs Markdown output
* Test for schema=pydantic_model
* Initial --schema CLI documentation
* Python docs for schema=
* Advanced plugin docs on schemas
2025-02-26 16:58:28 -08:00
Simon Willison
7bf1ea665e Made load_conversation() async_ aware, closes #742 2025-02-16 20:19:38 -08:00
Simon Willison
24b250506b Better __repr__ and __str__ for conversation and model
Inspired by work on #752
2025-02-16 15:45:08 -08:00
Simon Willison
2066397aae llm logs --prompts is now -s/--short - also supports --usage
Refs #736, closes #756
2025-02-16 15:12:28 -08:00
Simon Willison
6c6b100f3e
KeyModel and AsyncKeyModel classes for models that taken keys (#753)
* New KeyModel and AsyncKeyModel classes for models that taken keys - closes #744
* llm prompt --key now uses new mechanism, including for async
* use new key mechanism in llm chat command
* Python API tests for llm.KeyModel and llm.AsyncKeyModel
* Python API docs for for prompt(... key="")
* Mention await model.prompt() takes other parameters, reorg sections
* Better title for the model tutorial
* Docs on writing model plugins that take a key
2025-02-16 14:38:51 -08:00
Simon Willison
20c18a716d -q multiple option for llm models and llm embed-models
Refs #748
2025-02-13 15:35:18 -08:00
Simon Willison
41d64a8f12
llm logs --prompts option (#737)
Closes #736
2025-02-02 12:03:01 -08:00
Simon Willison
656d8fa3c4
--xl/--extract-last flag for prompt and log list commands (#718)
Closes #717
2025-01-24 10:52:46 -08:00
Csaba Henk
88a8cfd9e4
llm logs -x/--extract option (#693)
* llm logs -x/--extract option
* Update docs/help.md for llm logs -x
* Added test for llm logs -x/--extract, refs #693
* llm logs -xr behaves same as llm logs -x
* -x/--extract in llm logging docs

---------

Co-authored-by: Simon Willison <swillison@gmail.com>
2025-01-10 15:53:04 -08:00
Simon Willison
b452effa09 llm models -q/--query option, closes #700 2025-01-09 11:37:33 -08:00
Simon Willison
67d4a99645 llm prompt -x/--extract option, closes #681 2024-12-19 06:40:05 -08:00
Simon Willison
571f4b2a4d
Fix for UTC warnings
Closes #672
2024-12-12 14:57:23 -08:00
Simon Willison
b6be09aa28 Fix get_models() and get_async_models() duplicates bug
Closes #667, refs #640
2024-12-05 13:44:07 -08:00
Simon Willison
f9af563df5 response.on_done() mechanism, closes #653 2024-12-01 15:47:23 -08:00
Simon Willison
c52cfee881 llm.get_models() and llm.get_async_models(), closes #640 2024-11-20 20:09:06 -08:00
Simon Willison
cfb10f4afd
Log input tokens, output tokens and token details (#642)
* Store input_tokens, output_tokens, token_details on Response, closes #610
* llm prompt -u/--usage option
* llm logs -u/--usage option
* Docs on tracking token usage in plugins
* OpenAI default plugin logs usage
2024-11-19 20:21:59 -08:00
Simon Willison
ba75c674cb
llm.get_async_model(), llm.AsyncModel base class and OpenAI async models (#613)
- https://github.com/simonw/llm/issues/507#issuecomment-2458639308

* register_model is now async aware

Refs https://github.com/simonw/llm/issues/507#issuecomment-2458658134

* Refactor Chat and AsyncChat to use _Shared base class

Refs https://github.com/simonw/llm/issues/507#issuecomment-2458692338

* fixed function name

* Fix for infinite loop

* Applied Black

* Ran cog

* Applied Black

* Add Response.from_row() classmethod back again

It does not matter that this is a blocking call, since it is a classmethod

* Made mypy happy with llm/models.py

* mypy fixes for openai_models.py

I am unhappy with this, had to duplicate some code.

* First test for AsyncModel

* Still have not quite got this working

* Fix for not loading plugins during tests, refs #626

* audio/wav not audio/wave, refs #603

* Black and mypy and ruff all happy

* Refactor to avoid generics

* Removed obsolete response() method

* Support text = await async_mock_model.prompt("hello")

* Initial docs for llm.get_async_model() and await model.prompt()

Refs #507

* Initial async model plugin creation docs

* duration_ms ANY to pass test

* llm models --async option

Refs https://github.com/simonw/llm/pull/613#issuecomment-2474724406

* Removed obsolete TypeVars

* Expanded register_models() docs for async

* await model.prompt() now returns AsyncResponse

Refs https://github.com/simonw/llm/pull/613#issuecomment-2475157822

---------

Co-authored-by: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
2024-11-13 17:51:00 -08:00
Simon Willison
12df1a3b2a Show attachment types in llm models --options, closes #612 2024-11-05 22:49:26 -08:00
Simon Willison
6deed8f976 get_model() improvement, get_default_model() / set_default_wodel() now documented
Refs #553
2024-08-18 17:37:31 -07:00
Simon Willison
a83421607a Switch default model to gpt-4o-mini (from gpt-3.5-turbo), refs #536 2024-07-18 11:57:19 -07:00
Simon Willison
964f4d9934 Fix for llm logs -q plus -m bug, closes #515 2024-06-16 14:35:38 -07:00
Simon Willison
fb63c92cd2 llm logs -r/--response option, closes #431 2024-03-04 13:29:07 -08:00
Simon Willison
8021e12aaa
Windows readline fix, plus run CI against macOS and Windows
* Run CI on Windows and macOS as well as Ubuntu, refs #407
* Use pyreadline3 on win32
* Back to fail-fast since we have a bigger matrix now
* Mark some tests as xfail on windows
2024-01-26 16:24:58 -08:00
Simon Willison
214fcaaf86
Upgrade to run against OpenAI >= 1.0
* strategy: fail-fast: false - to help see all errors
* Apply latest Black

Refs #325
2024-01-25 22:00:44 -08:00
Simon Willison
b4ec54ef19 NotImplementedError for system prompts with OpenAI completion models, refs #284
Signed-off-by: Simon Willison <swillison@gmail.com>
2023-09-18 22:51:22 -07:00
Simon Willison
f76b2120e4 Revert "Handle system prompts for completion models, refs #284"
This reverts commit 4eed871cf1.

Decesion made in #288
2023-09-18 22:44:38 -07:00
Simon Willison
4eed871cf1 Handle system prompts for completion models, refs #284 2023-09-18 22:36:38 -07:00
Simon Willison
fcff36c6bc completion: true to register completion models, refs #284 2023-09-18 22:17:26 -07:00
Simon Willison
4fea46113f logprobs support for OpenAI completion models, refs #284 2023-09-18 22:04:28 -07:00
Simon Willison
2b504279d9 Test for OpenAI chat streaming, closes #287 2023-09-18 21:27:36 -07:00
Simon Willison
4d18da4e11 Bump default gpt-3.5-turbo-instruct max tokens to 256, refs #284 2023-09-18 20:29:39 -07:00
Simon Willison
4d46ebaa32 OpenAI completion models including gpt-3.5-turbo-instruct, refs #284 2023-09-18 18:34:32 -07:00
Simon Willison
6c43948325 llm.user_dir() creates directory if needed, closes #275
Would have fixed this bug too:
- https://github.com/simonw/llm-sentence-transformers/issues/9
2023-09-13 15:58:18 -07:00
Simon Willison
b30f6894f7 Fixed bug if LLM directory does not exist, closes #193 2023-08-31 20:26:05 -07:00
Simon Willison
c8e9565f47 Combine piped and argument prompts, closes #153 2023-08-20 22:57:29 -07:00