> they don't provide embedings, but storage and query engines for embeddings, so still very relevant
But you don't need any of the chain of: extract data, calculate embeddings, store data indexed by embeddings, detect need to retrieve data by embeddings and stuff it into LLM context along with your prompt if you use OpenAI's Assistants API, which, in addition to letting you store your own prompts and manage associated threads, also lets you upload data for it to extract, store, and use for RAG on the level of either a defined Assistant or a particular conversation (Thread.)
As in, use an existing search and call it via 'function calling' as part of the assistants routine - rather than uploading documents to the assistant API?
I mean embeddingsDB startups don't provide embeddings. They provide databases which allows to store and query computed embeddings (e.g. computed by ChatGPT), so they are complimentary services.
Yeah I still see a chat bot being able to look for related information in a database as useful. But I see it as just one of many tools a good chat experience will require. 128k context means for me there other applications to explore and larger tasks to accomplish with fewer api requests. Better chat history and context not getting lost
they don't provide embedings, but storage and query engines for embeddings, so still very relevant
> - file processing startups -> don't need to process files anymore
curious what is that exactly?..
> - vertical ai agent startups -> GPT marketplace
sure, those startups will be selling their agents on marketplace