Oracle Private AI Services Container FAQ

The Oracle Private AI Services Container gives Oracle AI Database customers a private, air-gap-capable, OpenAI-style inference layer that keeps embedding work off the database server while still fitting naturally into Oracle-native workflows. Protect Oracle AI Database performance and keep AI inference inside your security boundary.

Frequently Asked Questions

  • General Questions

    • Which services are provided by the container?
    • How do you communicate with the container?
    • How much memory is needed?
    • How many CPU cores are needed?
    • How much disk space is needed?
    • Are GPUs needed?
    • Is the container free to use?
    • Which 3rd party licenses does the Private AI Services Container use?
  • Configuration Questions

    • How do you control the log level of the container?
    • Which TCP ports does the container use?
    • Is the config.json file always needed?
  • Embedding Model Questions

    • Which vector embedding models ship with the container?
    • How do I use other embedding models in the container?
    • Which embedding models work with the container?
    • Which vector embedding model should I use?
    • Can the container create OpenAI-compatible vectors?
  • Deployment Questions

    • Which container runtimes are supported?
    • Can more than one container be used?
    • How is high availability supported?
    • Does the container run on the Oracle AI Database server?
    • Where can the container be run?
    • Can the container run in an air gapped environment?
    • What Linux software is required for the container to run on?
    • How many concurrent users are supported?
  • Security Questions

    • Is SSL supported?
    • How are the passwords stored on the container?
    • How are API Keys used?
    • Is user data stored in the container?
    • What security does the container use?
  • Vector Index Service Questions

    • Is a GPU required?
    • Will it work on my GeFORCE RTX 3070?
    • Will it work on an NVIDIA DGX Spark or Jetsen?
    • How much VRAM is needed?
    • How much RAM is needed?
    • Which version of the NVIDIA driers are needed?
    • Which version of the NVIDIA Container Toolkit is needed?
    • How do I choose which of my installed GPUs to use?
    • Why is the container not starting?
    • How can I check whether the container is running?
    • Will this work with HNSW indexes which are local or have include columns
    • Will this work with IVF indexes
    • Will this work with RAC, Exadata or sharded databases?
    • What parameters do I need in my CREATE VECTOR INDEX statement?
    • Why can the database not communicate with the container?