Truncation

This commit is contained in:
Roy Han 2024-07-01 11:59:44 -07:00
parent 1daac52651
commit 9c32b6b9ed
2 changed files with 1 additions and 20 deletions

View File

@ -996,25 +996,6 @@ If `stream` is set to `false`, then the response is a single JSON object:
{ "status": "success" }
```
```shell
POST /api/embed
```
# Parameters
- `model`
- `input`: string or array of strings
- cap on string size and array size
# Response
- `embeddings`: array of embeddings (array of floats)
- normalized? definitely normalize for openai
- truncate flag defaults true, otherwise false if doesn't fit
- deprecate /api/embeddings later
## Generate Embeddings
```shell

View File

@ -402,7 +402,7 @@ func (s *Server) EmbedHandler(c *gin.Context) {
}
if len(tokens) > opts.NumCtx {
tokens = tokens[len(tokens)-opts.NumCtx:]
tokens = tokens[:opts.NumCtx]
return runner.llama.Detokenize(c.Request.Context(), tokens)
}