Unification of capabilities. Now we have considerably simplified the interface of the /embeddings endpoint by merging the 5 separate fashions proven above (
code-search-code) right into a single new mannequin. This single illustration performs higher than our earlier embedding fashions throughout a various set of textual content search, sentence similarity, and code search benchmarks.
Longer context. The context size of the brand new mannequin is elevated by an element of 4, from 2048 to 8192, making it extra handy to work with lengthy paperwork.
Smaller embedding dimension. The brand new embeddings have solely 1536 dimensions, one-eighth the dimensions of
davinci-001 embeddings, making the brand new embeddings more economical in working with vector databases.
Lowered value. Now we have decreased the value of recent embedding fashions by 90% in comparison with previous fashions of the identical dimension. The brand new mannequin achieves higher or comparable efficiency because the previous Davinci fashions at a 99.8% decrease value.
General, the brand new embedding mannequin is a way more highly effective device for pure language processing and code duties. We’re excited to see how our prospects will use it to create much more succesful functions of their respective fields.