Truncation
This commit is contained in:
parent
1daac52651
commit
9c32b6b9ed
19
docs/api.md
19
docs/api.md
@ -996,25 +996,6 @@ If `stream` is set to `false`, then the response is a single JSON object:
|
||||
{ "status": "success" }
|
||||
```
|
||||
|
||||
```shell
|
||||
POST /api/embed
|
||||
```
|
||||
|
||||
# Parameters
|
||||
- `model`
|
||||
- `input`: string or array of strings
|
||||
- cap on string size and array size
|
||||
|
||||
# Response
|
||||
- `embeddings`: array of embeddings (array of floats)
|
||||
- normalized? definitely normalize for openai
|
||||
|
||||
- truncate flag defaults true, otherwise false if doesn't fit
|
||||
|
||||
- deprecate /api/embeddings later
|
||||
|
||||
|
||||
|
||||
## Generate Embeddings
|
||||
|
||||
```shell
|
||||
|
@ -402,7 +402,7 @@ func (s *Server) EmbedHandler(c *gin.Context) {
|
||||
}
|
||||
|
||||
if len(tokens) > opts.NumCtx {
|
||||
tokens = tokens[len(tokens)-opts.NumCtx:]
|
||||
tokens = tokens[:opts.NumCtx]
|
||||
return runner.llama.Detokenize(c.Request.Context(), tokens)
|
||||
}
|
||||
|
||||
|
Loading…
x
Reference in New Issue
Block a user