Supported LLMs
Last updated
Last updated
Valence supports a number of capable and versatile LLMs out-of-the box. You can use any of these models with the runLlm
function defined in Inference API
You can find the list below:
Model | Model ID |
---|---|
Llama-3-8B-Instruct
meta-llama/Meta-Llama-3-8B-Instruct