mirror of
https://github.com/Hopiu/llm.git
synced 2026-03-16 20:50:25 +00:00
Update docs to no longer mention PaLM
!stable-docs
This commit is contained in:
parent
64f9f2ef52
commit
e46cb7e761
3 changed files with 20 additions and 13 deletions
|
|
@ -31,7 +31,6 @@ These plugins can be used to interact with remotely hosted models via their API:
|
|||
- **[llm-anyscale-endpoints](https://github.com/simonw/llm-anyscale-endpoints)** supports models hosted on the [Anyscale Endpoints](https://app.endpoints.anyscale.com/) platform, including Llama 2 70B.
|
||||
- **[llm-replicate](https://github.com/simonw/llm-replicate)** adds support for remote models hosted on [Replicate](https://replicate.com/), including Llama 2 from Meta AI.
|
||||
- **[llm-fireworks](https://github.com/simonw/llm-fireworks)** supports models hosted by [Fireworks AI](https://fireworks.ai/).
|
||||
- **[llm-palm](https://github.com/simonw/llm-palm)** adds support for Google's [PaLM 2 model](https://developers.generativeai.google/).
|
||||
- **[llm-openrouter](https://github.com/simonw/llm-openrouter)** provides access to models hosted on [OpenRouter](https://openrouter.ai/).
|
||||
- **[llm-cohere](https://github.com/Accudio/llm-cohere)** by Alistair Shepherd provides `cohere-generate` and `cohere-summarize` API models, powered by [Cohere](https://cohere.com/).
|
||||
- **[llm-bedrock](https://github.com/simonw/llm-bedrock)** adds support for Nova by Amazon via Amazon Bedrock.
|
||||
|
|
|
|||
|
|
@ -38,34 +38,42 @@ llm plugins
|
|||
```json
|
||||
[
|
||||
{
|
||||
"name": "llm-mpt30b",
|
||||
"name": "llm-anthropic",
|
||||
"hooks": [
|
||||
"register_models"
|
||||
],
|
||||
"version": "0.11"
|
||||
},
|
||||
{
|
||||
"name": "llm-gguf",
|
||||
"hooks": [
|
||||
"register_commands",
|
||||
"register_models"
|
||||
],
|
||||
"version": "0.1"
|
||||
"version": "0.1a0"
|
||||
},
|
||||
{
|
||||
"name": "llm-palm",
|
||||
"name": "llm-clip",
|
||||
"hooks": [
|
||||
"register_commands",
|
||||
"register_models"
|
||||
"register_embedding_models"
|
||||
],
|
||||
"version": "0.1"
|
||||
},
|
||||
{
|
||||
"name": "llm.default_plugins.openai_models",
|
||||
"name": "llm-cmd",
|
||||
"hooks": [
|
||||
"register_commands",
|
||||
"register_models"
|
||||
]
|
||||
"register_commands"
|
||||
],
|
||||
"version": "0.2a0"
|
||||
},
|
||||
{
|
||||
"name": "llm-gpt4all",
|
||||
"name": "llm-gemini",
|
||||
"hooks": [
|
||||
"register_embedding_models",
|
||||
"register_models"
|
||||
],
|
||||
"version": "0.1"
|
||||
"version": "0.3"
|
||||
}
|
||||
]
|
||||
```
|
||||
|
|
|
|||
4
setup.py
4
setup.py
|
|
@ -15,8 +15,8 @@ def get_long_description():
|
|||
setup(
|
||||
name="llm",
|
||||
description=(
|
||||
"A CLI utility and Python library for interacting with Large Language Models, "
|
||||
"including OpenAI, PaLM and local models installed on your own machine."
|
||||
"CLI utility and Python library for interacting with Large Language Models from "
|
||||
"organizations like OpenAI, Anthropic and Gemini plus local models installed on your own machine."
|
||||
),
|
||||
long_description=get_long_description(),
|
||||
long_description_content_type="text/markdown",
|
||||
|
|
|
|||
Loading…
Reference in a new issue