Oracle Data Platform for Financial Services

Fraud prevention and anti–money laundering

Minimize risk while delivering seamless customer service

Financial fraud poses a major challenge for the financial services industry. Not only does it come in many different forms, but it’s often difficult to detect due to the complexity of the relationships between entities and hidden patterns. And, once detected, financial institutions must notify customers of fraudulent activity in real time and take immediate action to stop it—for example, by blocking the customer’s credit card.

The financial services industry is also regulated and must report anti–money laundering (AML) activities and complete due diligence on their customers using Know Your Customer (or KYC) processes. This often requires analyzing data across products, markets, and geographies to identify relationships and patterns for AML.

Conceptually, money laundering is simple: Dirty money is passed around, blended with legitimate funds, and then turned into hard assets. In reality, it’s far more complicated, relying on a long, complex series of valid transfers between accounts created using synthetic (often stolen) identities and often using similar information, such as email and street addresses. In short, it involves a vast amount of data, which is why a unified data platform that supports advanced analytic techniques, such as graph analysis, is essential for AML programs.

Protect both customers and institutions with machine learning

The financial services industry continues to be highly monitored and regulated, and few areas have seen a greater increase in regulatory focus than anti–money laundering and counter-terrorist financing activities. Driven by vast criminal networks, financial fraud is a sophisticated and growing challenge requiring anti–money laundering solutions that provide insight across the enterprise and the entire globe.

The following architecture demonstrates how Oracle components and capabilities, including advanced analytics and machine learning, can be combined to create a data platform that covers the entire data analytics lifecycle and delivers the insights AML teams need to identify the anomalous patterns in behavior that can be indicative of fraudulent activity.

Fraud prevention and AML diagram, description below

This image shows how Oracle Data Platform for financial services can be used to support fraud prevention and AML activities. The platform includes the following five pillars:

  • Data Sources, Discovery
  • Connect, Ingest, Transform
  • Persist, Curate, Create
  • Analyze, Learn, Predict
  • Measure. Act

The Data Sources, Discovery pillar includes two categories of data.

Business record (first-party) data comprises credit card transactions, account information, and ATM event streams.

Third-party data includes social feeds.

The Connect, Ingest, Transform pillar comprises four capabilities.

Bulk transfer uses OCI FastConnect, OCI Data Transfer, MFT, and OCI CLI.

Batch ingestion uses OCI Data Integration, Oracle Data Integrator, and DB tools.

Change data capture uses OCI GoldenGate and Oracle Data Integrator.

Streaming ingest uses OCI Streaming and Kafka Connect.

All four capabilities connect unidirectionally into the cloud storage/data lake capability within the Persist, Curate, Create pillar.

Additionally, streaming ingest is connected to stream processing within the Analyze, Learn, Predict pillar.

The Persist, Curate, Create pillar comprises four capabilities.

The serving data store uses Autonomous Data Warehouse and Exadata Cloud Service.

Cloud storage/data lake uses OCI Object Storage.

Batch processing uses OCI Data Flow.

Governance uses OCI Data Catalog.

These capabilities are connected within the pillar. Cloud storage/data lake is unidirectionally connected to the serving data store; it is also bidirectionally connected to batch processing.

One capability connects into the Analyze, Learn, Predict pillar: The serving data store connects unidirectionally to the analytics and visualization, AI services, and machine learning capabilities and bidirectionally to the streaming analytics capability.

The Analyze, Learn, Predict pillar comprises four capabilities.

Analytics and visualization uses Oracle Analytics Cloud, GraphStudio, and ISVs.

AI services uses OCI Anomaly Detection, OCI Forecasting, OCI Language

Machine learning uses OCI Data Science and Oracle Machine Learning Notebooks.

Streaming analytics uses OCI GoldenGate Stream Analytics.

The Measure, Act pillar comprises three consumers: dashboards and reports, applications, and machine learning models.

People and partners comprise Geographic Heat-Maps, Regulatory Reporting, High Risk Entity Predictive Analytics, Fraud Trends.

Applications comprises Transaction Anomaly Detection, Text Mining and Natural Language Processing (NLP).

Machine learning models comprises Network Analysis, Behavior Analysis, Watchlist Screening.

The three central pillars—Ingest, Transform; Persist, Curate, Create; and Analyze, Learn, Predict—are supported by infrastructure, network, security, and IAM.


There are three main ways to inject data into an architecture to enable financial services organizations to identify potentially fraudulent activity.

  • To start, we need data from transactional systems and core banking applications. This data can be enriched with customer data from third-party sources, which could include unstructured data from social media, for example. Frequent real-time or near real-time extracts requiring change data capture are common, and data is regularly ingested from transactional, risk, and customer management systems using Oracle Cloud Infrastructure (OCI) GoldenGate. OCI GoldenGate is also a critical component of evolving data mesh architectures where “data products” are managed via enterprise data ledgers and polyglot data streams that perform continuous transform and load processes (rather than the batch ingest and extract, transform, and load processes used in monolithic architectures).
  • We can now use streaming ingest to ingest data from IoT sensors, web pipelines, log files, point-of-sale devices, ATMs, social media, and other data sources in real time via OCI Streaming/Kafka. This streamed data (events) is ingested and some basic transformations/aggregations are performed before the data is stored in cloud storage. In parallel with ingestion, we can filter, aggregate, correlate, and analyze high volumes of data from multiple sources in real time using streaming analytics. This not only helps financial institutions detect business threats and risks—for example, suspicious transactions from an ATM, such as multiple repeat transactions— but it also provides insights into their overall fraud prevention efficiency. Correlating events and identified patterns can be fed back (manually) for a data science examination of the raw data. Additionally, events can be generated to trigger actions, such as notifying customers about potential fraud via email or SMS or blocking compromised debit cards. Oracle GoldenGate Stream Analytics is an in-memory technology that performs real-time analytical computations on streaming data.
  • While real-time needs are evolving, the most common extract from core banking, customer, and financial systems is a batch ingestion using an extract, transform, and load process. Batch ingestion is used to import data from systems that can’t support streaming ingestion (for example, older mainframe systems). For anti–money laundering and Know Your Customer processes, data is sourced from different operational systems, such as current and credit account transaction processing systems, and third-party data feeds providing customer intelligence. The data is sourced across products and geographies. Batch ingestions can be frequent, as often as every 10 or 15 minutes, but they are still bulk in nature as groups of transactions are extracted and processed rather than individual transactions. OCI offers different services to handle batch ingestion, such as the native OCI Data Integration service or Oracle Data Integrator running on an OCI Compute instance. Depending on the volumes and data types, data can be loaded into object storage or loaded directly into a structured relational database for persistent storage.

Data persistence and processing is built on three (optionally four) components.

  • Ingested raw data is stored in cloud storage for algorithmic purposes; we use OCI Object Storage as the primary data persistence tier. Spark in OCI Data Flow is the primary batch processing engine for data such as transactional, location, application, and geo-mapping data. Batch processing involves several activities, including basic noise treatment, missing data management, and filtering based on defined outbound datasets. Results are written back to various layers of object storage or to a persistent relational repository based on the processing needed and the data types used.
  • These processed datasets are returned to cloud storage for onward persistence, curation, and analysis and ultimately for loading in optimized form to the serving data store, provided here by Oracle Autonomous Database. Data is now persisted in optimized relational form for curation and query performance. Alternatively, depending on architectural preference, this can be accomplished with Oracle Big Data Service as a managed Hadoop cluster. In this use case, all the data needed to train the machine learning models is accessed in raw form from object storage. To train the models, historical patterns are combined with transaction-level records to identify and label potential risks. Combining these datasets with others, such as device data and geospatial data, lets us apply data science techniques to refine existing models and develop new ones to better predict potential frauds. This type of persistence can also be used to store data for schemas that are part of the data stores accessed via external tables and hybrid partitions.

The ability to analyze, predict, and act is built on three technology approaches.

  • Analytics and visualization services, such as Oracle Analytics Cloud, deliver analytics based on the curated data from the serving data store. This includes descriptive analytics (describes current fraud identification trends and flagged activity with histograms and charts), predictive analytics, such as time series analysis (predicts future patterns, identifies trends, and determines the probability of uncertain outcomes), and prescriptive analytics (proposes suitable actions leading to optimal decision-making). These analytics can be used to answer questions such as: How does actual flagged fraud this period compare to previous periods?
  • Alongside advanced analytics, machine learning models are developed, trained, and deployed. These trained models can be run on both current and historical transactional data to detect money laundering by matching patterns of transactions and behaviors, and the results can be persisted back to the serving layer and reported using analytics tools such as Oracle Analytics Cloud. To optimize model training, the model and data can also be fed into machine learning systems, such as OCI Data Science, to further train the models for more effective anti–money laundering pattern detection before promoting them. These models can be accessed via APIs, deployed within the serving data store, or embedded as part of the OCI GoldenGate streaming analytics pipeline.
  • Additionally, we can use the advanced capabilities of cloud native artificial intelligence services.
    • OCI Anomaly Detection is an artificial intelligence service that makes it easy to build business-specific anomaly detection models that flag critical incidents, speeding up detection and resolution. In this use case, we would deploy these models to detect fraud during the lifecycle of a transaction; during audits; in specific contexts, for example, based on the vendor, merchant, or type of transaction; and in many other scenarios. OCI Anomaly Detection can identify all these types of fraud by using historical data and building an appropriate anomaly detection model. For example, if the dataset includes the type of transaction with the amount, location (latitude and longitude), vendor name, and other details, OCI Anomaly Detection can identify if the fraud is related to the transaction amount, the transaction account, the location where the transaction occurred, or the vendor who filed the transaction.
    • OCI Forecasting can be used to forecast transaction metrics, such as the number of transactions, transaction amounts, and so on, for the next day, week, or months as a function of current metrics and influencing market conditions. These forecasts can be used for planning and to set a baseline expectation to use in safeguarding against money laundering and other fraud.
    • OCI Language and OCI Vision can ingest documents and text that can help enrich the data for fraud detection and AML activities.
  • Data governance is another critical component. This is delivered by OCI Data Catalog, a free service providing data governance and metadata management (for both technical and business metadata) for all the data sources in the data lakehouse ecosystem. OCI Data Catalog is also a critical component for queries from Oracle Autonomous Data Warehouse to OCI Object Storage as it provides a way to quickly locate data regardless of its storage method. This allows end users, developers, and data scientists to use a common access language (SQL) across all the persisted data stores in the architecture.
  • Finally, our now curated, tested, high-quality, and governed data and models can be exposed as a data product (API) within a data mesh architecture for distribution across the financial services organization.

Improve fraud prevention and AML activities with the right data platform

Oracle Data Platform can help your organization detect money laundering more effectively, boost the accuracy and efficiency of financial crime investigations, and streamline reporting processes to keep compliance costs down.

Related resources

Начало работы с современной платформой данных Oracle

Более 20 бесплатных облачных служб Always Free в 30-дневной пробной версии

Oracle предлагает бесплатную пробную версию без ограничений по времени для более чем 20 сервисов, таких как Autonomous Database и Arm Compute и Storage, а также бонусы на 300 долларов США для пробного использования дополнительных облачных сервисов. Узнайте подробности и зарегистрируйтесь бесплатно уже сегодня.

  • Что предлагается в рамках Oracle Cloud Free Tier?

    • 2 автономные базы данных, объемом 20 ГБ каждая
    • Виртуальные машины AMD и Arm Compute
    • Общее блочное хранилище объемом 200 ГБ
    • Объектное хранилище на 10 ГБ
    • 10 ТБ исходящих данных в месяц
    • Более 10 бесплатных сервисов Always Free
    • Бонус в 300 долларов США сроком на месяц и даже больше

Учитесь с помощью пошаговых инструкций

Ознакомьтесь с широким спектром сервисов OCI с помощью учебных пособий и тренингов. Независимо от того, являетесь Вы разработчиком, администратором или аналитиком, мы поможем Вам понять, как работает OCI. Многие практические занятия проходят на уровне Oracle Cloud Free Tier или на бесплатной платформе для практических занятий Oracle.

  • Начало работы с базовыми сервисами OCI

    Практические занятия этого семинара охватывают введение в основные сервисы Oracle Cloud Infrastructure (OCI), включая виртуальные облачные сети (VCN), а также сервисы вычислительных ресурсов и хранения.

    Начать практическое занятие «Базовые сервисы OCI»
  • Быстрый запуск Autonomous Database

    На этом семинаре Вы ознакомитесь с пошаговыми инструкциями по началу работы с Oracle Autonomous Database.

    Начать практическое занятие Autonomous Database
  • Создание приложения из электронной таблицы

    В рамках этого практического занятия Вы загрузите электронную таблицу в таблицу базы данных Oracle, а затем создадите приложение на основе этой новой таблицы.

    Начать этот тренинг
  • Развертывание приложения HA в OCI

    На этом практическом занятии Вы развернете веб-серверы на двух вычислительных экземплярах в Oracle Cloud Infrastructure (OCI), настроенных в режиме высокой доступности с помощью балансировщика нагрузки.

    Начать практическое занятие «Приложение с высокой доступностью»

Изучите более 150 лучших практик

Посмотрите, как архитекторы и заказчики Oracle развертывают различные нагрузки: от корпоративных приложений до высокопроизводительных вычислений и от микросервисов до озер данных. Ознакомьтесь с лучшими практиками, узнайте много нового от архитекторов-заказчиков в нашей серии видео Built & Deployed, а также разверните множество нагрузок либо благодаря возможности «нажмите, чтобы развернуть», либо самостоятельно с помощью репозитория Oracle GitHub.

Популярные архитектуры

  • Apache Tomcat и сервис MySQL Database
  • Oracle Weblogic в Kubernetes и Jenkins
  • Среды машинного обучения и искусственного интеллекта
  • Tomcat on Arm и Oracle Autonomous Database
  • Анализ журналов со стеком ELK
  • HPC и OpenFOAM

Узнайте, сколько можно сэкономить благодаря возможностям OCI

Ценообразование Oracle Cloud построено на принципах простоты и постоянства с поддержкой широкого спектра сценариев использования. Чтобы оценить низкую ставку, откройте калькулятор затрат и настройте сервисы в соответствии с Вашими потребностями.

Почувствуйте разницу:

  • 1/4 исходящих затрат на пропускную способность
  • 3-кратное соотношение «цена-производительность» для вычислений
  • Одинаковая низкая цена в каждом регионе
  • Низкие цены без долгосрочных обязательств

Связаться с отделом продаж

Хотите узнать больше об Oracle Cloud Infrastructure? Позвольте одному из экспертов Oracle помочь.

  • Они могут ответить на такие вопросы, как:

    • Какие нагрузки лучше всего выполняются в OCI?
    • Как получить максимальную отдачу от инвестиций в Oracle?
    • Чем OCI отличается от облачных вычислений других поставщиков?
    • Как может OCI помочь Вам в достижении целей по IaaS и PaaS?