Unification of capabilities. Now we have considerably simplified the interface of the /embeddings endpoint by merging the 5 separate fashions proven above (
code-search-code) right into a single new mannequin. This single illustration performs higher than our earlier embedding fashions throughout a various set of textual content search, sentence similarity, and code search benchmarks.
Longer context. The context size of the brand new mannequin is elevated by an element of 4, from 2048 to 8192, making it extra handy to work with lengthy paperwork.
Smaller embedding measurement. The brand new embeddings have solely 1536 dimensions, one-eighth the scale of
davinci-001 embeddings, making the brand new embeddings less expensive in working with vector databases.
Lowered worth. Now we have lowered the value of recent embedding fashions by 90% in comparison with outdated fashions of the identical measurement. The brand new mannequin achieves higher or comparable efficiency because the outdated Davinci fashions at a 99.8% decrease worth.
Total, the brand new embedding mannequin is a way more highly effective software for pure language processing and code duties. We’re excited to see how our prospects will use it to create much more succesful functions of their respective fields.