Categories: FAANG

New and Improved Embedding Model

We are excited to announce a new embedding model which is significantly more capable, cost effective, and simpler to use. The new model, text-embedding-ada-002, replaces five separate models for text search, text similarity, and code search, and outperforms our previous most capable model, Davinci, at most tasks, while being priced 99.8% lower.

Read documentation

Embeddings are numerical representations of concepts converted to number sequences, which make it easy for computers to understand the relationships between those concepts. Since the initial launch of the OpenAI /embeddings endpoint, many applications have incorporated embeddings to personalize, recommend, and search content.







You can query the /embeddings endpoint for the new model with two lines of code using our OpenAI Python Library, just like you could with previous models:

import openai
response = openai.Embedding.create(
  input="porcine pals say",
  model="text-embedding-ada-002"
)

print(response)
{
  "data": [
    {
      "embedding": [
        -0.0108,
        -0.0107,
        0.0323,
        ...
        -0.0114
      ],
      "index": 0,
      "object": "embedding"
    }
  ],
  "model": "text-embedding-ada-002",
  "object": "list"
}

Model Improvements

Stronger performance. text-embedding-ada-002 outperforms all the old embedding models on text search, code search, and sentence similarity tasks and gets comparable performance on text classification. For each task category, we evaluate the models on the datasets used in old embeddings.





Unification of capabilities. We have significantly simplified the interface of the /embeddings endpoint by merging the five separate models shown above (text-similarity, text-search-query, text-search-doc, code-search-text and code-search-code) into a single new model. This single representation performs better than our previous embedding models across a diverse set of text search, sentence similarity, and code search benchmarks.

Longer context. The context length of the new model is increased by a factor of four, from 2048 to 8192, making it more convenient to work with long documents.

Smaller embedding size. The new embeddings have only 1536 dimensions, one-eighth the size of davinci-001 embeddings, making the new embeddings more cost effective in working with vector databases.

Reduced price. We have reduced the price of new embedding models by 90% compared to old models of the same size. The new model achieves better or similar performance as the old Davinci models at a 99.8% lower price.

Overall, the new embedding model is a much more powerful tool for natural language processing and code tasks. We are excited to see how our customers will use it to create even more capable applications in their respective fields.

Limitations

The new text-embedding-ada-002 model is not outperforming text-similarity-davinci-001 on the SentEval linear probing classification benchmark. For tasks that require training a light-weighted linear layer on top of embedding vectors for classification prediction, we suggest comparing the new model to text-similarity-davinci-001 and choosing whichever model gives optimal performance.

Check the Limitations & Risks section in the embeddings documentation for general limitations of our embedding models.

Examples of Embeddings API in Action

Kalendar AI is a sales outreach product that uses embeddings to match the right sales pitch to the right customers out of a dataset containing 340M profiles. This automation relies on similarity between embeddings of customer profiles and sale pitches to rank up most suitable matches, eliminating 40–56% of unwanted targeting compared to their old approach.

Notion, the online workspace company, will use OpenAI’s new embeddings to improve Notion search beyond today’s keyword matching systems.


Read documentation

var initRotate = function () { var rotates = document.querySelectorAll('.js-rotate'); if (!rotates.length) return; // for each set of rotates rotates.forEach(function (r) { // move first child to end every n seconds window.setInterval(function(){ moveToEnd(r, r.firstElementChild); }, 1500); }); }; var moveToEnd = function (parent, child) { parent.removeChild(child); parent.appendChild(child); // append to parent };

var initShowMore = function () { var showmores = document.querySelectorAll('.js-showmore'); showmores.forEach(function (e) { e.addEventListener('click', function () { var showmoreparent = this.parentElement; if (!showmoreparent) return; var more = showmoreparent.querySelector('.js-more'); if (!more) return; more.style.display = 'block'; this.style.display = 'none'; }); }); };

// toggle function var toggle = function (whichIds, otherIds) { for (var i = 0; i < whichIds.length; i++) { var whichId = whichIds[i]; var whichEl = document.getElementById(whichId); if (!whichEl) return; whichEl.style.display = 'block'; } for (var i = 0; i < otherIds.length; i++) { var otherId = otherIds[i]; var otherEl = document.getElementById(otherId); if (!otherEl) return; otherEl.style.display = 'none'; } }; // togglers var initToggler = function () { var togglers = document.querySelectorAll('.js-toggler'); if (!togglers.length) return; for (var i = 0; i < togglers.length; i++) { var toggler = togglers[i]; toggler.addEventListener('click', function (e) { removeActiveTogglers(this.parentElement.querySelectorAll('.js-toggler')); addActiveToggler(this); }); } }; var addActiveToggler = function (el) { el.classList.add('active'); }; var removeActiveTogglers = function (els) { els.forEach(function (el) { el.classList.remove('active'); }); }; // init document.addEventListener('DOMContentLoaded', function () { initRotate(); initShowMore(); initToggler(); });

const customWidth = function (selector) { return (new Library).Generators.observe(function(change) { var width = change(document.querySelector(selector).clientWidth); function resized() { var w = document.querySelector(selector).clientWidth; if (w !== width) change(width = w); } window.addEventListener("resize", resized); return function() { window.removeEventListener("resize", resized); }; }); };

const tables_renders = { "tsTable": "#ts-table", "csTable": "#cs-table", "ssTable": "#ss-table", "tcTable": "#tc-table", }; new Runtime(Object.assign(new Library, {width: customWidth("#ts-table")})).module(notebook_tables, name => { const selector = tables_renders[name]; if (selector) { // key exists return new Inspector(document.querySelector(selector)); } else { return true; } });


Acknowledgments

Thanks to the following for their contributions to this release:
Chris Hallacy, Sherwin Wu, Jessica Shieh, Juston Forte, Aliisa Rosenthal, Katie Mayer

Thanks to the following for their feedback on this post:
Peter Welinder, Logan Kilpatrick, Joannne Jang, Fraser Kelton, Justin Jay Wang, Ruby Chen

AI Generated Robotic Content

Recent Posts

Workflow upscale/magnify video from Sora with Wan , based on cseti007

📦 : https://github.com/lovisdotio/workflow-magnify-upscale-video-comfyui-lovis I did this ComfyUI workflow for Sora 2 upscaling 🚀 ( or…

24 seconds ago

The Complete Guide to Pydantic for Python Developers

Python's flexibility with data types is convenient when coding, but it can lead to runtime…

30 seconds ago

Rooms from Motion: Un-posed Indoor 3D Object Detection as Localization and Mapping

We revisit scene-level 3D object detection as the output of an object-centric framework capable of…

39 seconds ago

Inside the AIPCon 8 Demos Transforming Manufacturing, Insurance, and Construction

Editor’s Note: This is the second in a two-part series highlighting demo sessions from AIPCon…

45 seconds ago

Responsible AI design in healthcare and life sciences

Generative AI has emerged as a transformative technology in healthcare, driving digital transformation in essential…

55 seconds ago

5 ad agencies used Gemini 2.5 Pro and gen media models to create an “impossible ad”

The conversation around generative AI in the enterprise is getting creative.  Since launching our popular…

57 seconds ago