diff --git a/docs/changelog.md b/docs/changelog.md index b37b0a5..6a0451e 100644 --- a/docs/changelog.md +++ b/docs/changelog.md @@ -1,10 +1,11 @@ # Changelog +(v0_5)= ## Unreleased -LLM now supports multiple language models, thanks to a new {ref}`plugins` hook for registering additional models. +LLM now supports **additional language models**, thanks to a new {ref}`plugins` mechanism for registering additional models. -Plugins are available for 19 models in addition to the default OpenAI ones. +Plugins are available for 19 models in addition to the default OpenAI ones: - [llm-gpt4all](https://github.com/simonw/llm-gpt4all) adds support for 17 models that can download and run on your own device, including Vicuna, Falcon and wizardLM. - [llm-mpt30b](https://github.com/simonw/llm-mpt30b) adds support for the MPT-30B model, a 19GB download. @@ -12,22 +13,26 @@ Plugins are available for 19 models in addition to the default OpenAI ones. A comprehensive tutorial, {ref}`writing a plugin to support a new model ` describes how to add new models by building plugins in detail. -- `llm "prompt" --save template` option for saving a prompt directly to a template. [#55](https://github.com/simonw/llm/issues/55) -- Prompt templates can now specify {ref}`default values ` for parameters. Thanks, Chris Mungall. [#57](https://github.com/simonw/llm/pull/57) -- New `-o/--option name value` syntax for setting options for models, such as temperature. Available options differ for different models. [#63](https://github.com/simonw/llm/issues/63) -- `llm models list --options` command for viewing all available model options. [#82](https://github.com/simonw/llm/issues/82) -- `llm -m 4-32k` alias for `gpt-4-32k`. -- `llm openai models` command to list all available OpenAI models from their API. [#70](https://github.com/simonw/llm/issues/70) -- `llm -s` is now a shortcut for `llm --system`. [#69](https://github.com/simonw/llm/issues/69) -- `llm models default MODEL_ID` to set a different model as the default to be used when `llm` is run without the `-m/--model` option. [#31](https://github.com/simonw/llm/issues/31) -- [Python API](https://llm.datasette.io/en/stable/python-api.html) documentation for using LLM models, including models from plugins, directly from Python. [#75](https://github.com/simonw/llm/issues/75) +### New features + +- {ref}`python-api` documentation for using LLM models, including models from plugins, directly from Python. [#75](https://github.com/simonw/llm/issues/75) - Messages are now logged to the database by default - no need to run the `llm init-db` command any more, which has been removed. Instead, you can toggle this behavior off using `llm logs off` or turn it on again using `llm logs on`. The `llm logs status` command shows the current status of the log database. If logging is turned off, passing `--log` to the `llm prompt` command will cause that prompt to be logged anyway. [#98](https://github.com/simonw/llm/issues/98) - New database schema for logged messages, with `conversations` and `responses` tables. If you have previously used the old `logs` table it will continue to exist but will no longer be written to. [#91](https://github.com/simonw/llm/issues/91) +- New `-o/--option name value` syntax for setting options for models, such as temperature. Available options differ for different models. [#63](https://github.com/simonw/llm/issues/63) +- `llm models list --options` command for viewing all available model options. [#82](https://github.com/simonw/llm/issues/82) +- `llm "prompt" --save template` option for saving a prompt directly to a template. [#55](https://github.com/simonw/llm/issues/55) +- Prompt templates can now specify {ref}`default values ` for parameters. Thanks, Chris Mungall. [#57](https://github.com/simonw/llm/pull/57) +- `llm openai models` command to list all available OpenAI models from their API. [#70](https://github.com/simonw/llm/issues/70) +- `llm models default MODEL_ID` to set a different model as the default to be used when `llm` is run without the `-m/--model` option. [#31](https://github.com/simonw/llm/issues/31) + +### Smaller improvements + +- `llm -s` is now a shortcut for `llm --system`. [#69](https://github.com/simonw/llm/issues/69) +- `llm -m 4-32k` alias for `gpt-4-32k`. - `llm install -e directory` command for installing a plugin from a local directory. -- [Comprehensive tutorial](https://llm.datasette.io/en/stable/plugins/tutorial-model-plugin.html) on writing plugins that implement new models. - The `LLM_USER_PATH` environment variable now controls the location of the directory in which LLM stores its data. This replaces the old `LLM_KEYS_PATH` and `LLM_LOG_PATH` and `LLM_TEMPLATES_PATH` variables. [#76](https://github.com/simonw/llm/issues/76) +- Documentation covering {ref}`plugin-utilities`. - Documentation site now uses Plausible for analytics. [#79](https://github.com/simonw/llm/issues/79) -- Documentation covering [utility functions for plugins](https://llm.datasette.io/en/latest/plugins/plugin-utilities.html). (v0_4_1)= ## 0.4.1 (2023-06-17) diff --git a/docs/plugins/plugin-utilities.md b/docs/plugins/plugin-utilities.md index 21cf4c4..44b5e13 100644 --- a/docs/plugins/plugin-utilities.md +++ b/docs/plugins/plugin-utilities.md @@ -1,3 +1,4 @@ +(plugin-utilities)= # Utility functions for plugins LLM provides some utility functions that may be useful to plugins. diff --git a/docs/python-api.md b/docs/python-api.md index 16816e7..a19a7b4 100644 --- a/docs/python-api.md +++ b/docs/python-api.md @@ -1,3 +1,4 @@ +(python-api)= # Python API LLM provides a Python API for executing prompts, in addition to the command-line interface.