New models available from Meta and Cohere for OCI Generative AI include
Dedicated AI clusters give you the ability to host foundational models on dedicated GPUs that are private to you. These clusters also provide you a stable throughput performance that's required for production use cases and can support hosting and fine-tuning workloads. OCI Generative AI gives you the ability to scale out your cluster with zero downtime to handle changes in volume call. Up to 50 custom, fine-tuned models can be hosted on the same dedicated hosting cluster as long as these fine-tuned models all share the same base foundational model.
OCI Generative AI will be integrated with LangChain, an open source framework that has emerged to develop new interfaces for generative AI applications powered by language models. LangChain makes it easy to swap out abstractions and components necessary to work with language models.
OCI Generative AI provides content moderation controls, endpoint model swap with zero downtime, and endpoints deactivation and activation capabilities. For each model endpoint, OCI Generative AI also captures a series of analytics including call statistics, tokens processed, error counts, etc.