A vector is a numerical representation of text, images, audio, or video which encodes the meaning of the data, not the underlying words or pixels.
A pretrained vector embedding model does an inference on an input (text, image, audio, or video) and generates a vector as the output. The vector is the values of the last hidden layer of the neural network after an inference for the input.
Oracle AI Vector Search supports up to 65,535 dimensions.
AI Vector Search supports the INT8, Float32, and Float64 formats.
Yes, you can create vectors inside the database via the vector_embedding() SQL function.
You choose Sentence-Transformer embedding models from Hugging Face, and then you can securely upload them to the database.
Yes, you can create vectors outside of the database using both commercial and open source models using either REST calls or local libraries.
The embedding models that have been tested include openai.com, cohere.com, Hugging Face Transformers, Sentence-Transformers, Transformers.js, and using the ONNX Runtime.
You can use create vectors using either CPUs or GPUs.
AI Vector Search should be able to work with any LLM.
So far, Llama2, Gemini, and PaLM 2 as well as OpenAI’s ChatGPT and LLMs developed by Cohere, Vertex AI, and Mistral AI have been tested.
More than 90 embedding models from OpenAI, Cohere, and ONNX Runtime have been tested, as have the Transformer, Sentence Transformer, Transformer.js, Xenova, and FastEmbed models.
Oracle AI Vector Search supports the Oracle AI Vector Search provider for LangChain.