Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: c4ai-command-r-v01 support #944

Open
0x77dev opened this issue Apr 6, 2024 · 2 comments
Open

feat: c4ai-command-r-v01 support #944

0x77dev opened this issue Apr 6, 2024 · 2 comments

Comments

@0x77dev
Copy link

0x77dev commented Apr 6, 2024

Feature request

Would be nice to have ability to run Command-R (CohereForAI/c4ai-command-r-v01) using OpenLLM

Motivation

No response

Other

vLLM backend already supports Command-R in v0.4.0: vllm-project/vllm#3330 (comment)

@0x77dev
Copy link
Author

0x77dev commented Apr 6, 2024

The current ghcr.io/bentoml/openllm:latest image (sha256:1860863091163a8e8cb1225c99d6e1b0735c11871e14e8d8424a22a5ad6742fa) shows an error:

ValueError: The checkpoint you are trying to load has a model type of `cohere`, which Transformers does not recognize. This may be due to a problem with the checkpoint or an outdated version of Transformers.

when doing this:

docker run --rm --gpus all -p 3000:3000 -it ghcr.io/bentoml/openllm start CohereForAI/c4ai-command-r-v01 --backend vllm

also when installing openllm[vllm] it brings 0.2.7 version of vLLM

Though vLLM version in main branch is 0.4.0:
https://github.com/bentoml/OpenLLM/blob/main/openllm-core/pyproject.toml#L83 and https://github.com/bentoml/OpenLLM/blob/main/tools/dependencies.py#L157

@GaetanBaert
Copy link

I think this should be the same prompting system, there is also CohereForAI/c4ai-command-r-plus available and it would be nice to be able to run it too.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants