Use-cases/Project
17 min read

Real-time Machine Learning: considerations based on Fraud Detection use case

When it comes to machine learning, most products are designed to work in batches, meaning they process data at fixed intervals rather than in real-time. This approach is often easier to manage, and in many cases, meets the needs of the business. However, there are situations where real-time machine learning is essential. In this article, we will explore why.

To understand why real-time machine learning is important, it's helpful to first understand the different paradigms of deployment for machine learning solutions.

Batch predictions

According to Databricks, batch deployment is the most common way of deploying machine learning models, accounting for approximately 80-90% of all deployments. This means running predictions from a model and saving them for later use. For live serving, results are saved to a low-latency database that will serve the predictions quickly. Alternatively, predictions can be stored in less performant data stores.

batch-prediction-getindata

At this stage, all predictions are precomputed periodically (e.g. daily) before a prediction request arises. Batch processing systems like Spark or MapReduce are used to process large amounts of data efficiently. By caching precomputed predictions, we also decouple computation from serving.

Typical use cases for batch predictions include churn prediction, customer segmentation, and sales forecasting. These use cases do not typically require real-time data.

Online predictions

Although real-time deployments represent a smaller share of the deployment landscape, they often involve high value tasks that provide significant business value. Real-time deployments can either use batch features or streaming features. It’s important to note that while features in real-time deployments might be processed online, the training of the machine learning models is generally done in batch mode. This is also to be distinguished from another area called Online ML, where not only inference data but also training data arrives in online mode, and specialized versions of ML algorithms are trained continuously. This last case, however, is not in the scope of this article.

Batch (static) features are pieces of information that rarely or never change, such as customer demographics (date of birth, gender, income) or product attributes (color, size, product category and images).

Streaming (online) features, on the other hand, are computed from streaming data and are based on real-time events. Examples include aggregate streamed events (such as impressions, clicks, likes, or purchases), social media data (such as sentiment analysis or trending topics), or sensor data from IoT devices.

Here are a few examples of real-time machine learning applications:

  • fraud detection
    In the banking industry, machine learning models can be used to predict fraudulent transactions in real-time, by analyzing patterns and identifying anomalies. This example is described in detail in the chapter below
  • product recommendations
    In the e-commerce industry, real-time product recommendations are extremely useful, since they can adapt to changes in user behavior and preferences over time. If a user browsing a website indicates a sudden interest in a particular category of products, we can update our recommendations to reflect this novel interest
  • dynamic pricing
    In the transportation industry, machine learning models can dynamically adjust the pricing based on real-time factors such as demand, weather or traffic conditions. When demand is unusually high, the price will increase accordingly
  • predictive maintenance
    In the manufacturing industry, companies can use real-time data from sensors and equipment to predict whether maintenance is needed, reducing downtime and improving efficiency. If a sensor records readings outside normal ranges, alerts can be sent to workers to inspect production pipelines.

Online predictions with batch features

While batch predictions are useful and sufficient for most use cases, they do have many limitations. For example, consider an ecommerce website that wants to predict the optimal discount for a user based on the products in their basket. The number of possible combinations of different products is unfeasible to precompute. Another example is chatbot, as it can be asked an endless variety of questions, often with a mixed order of words. The combinations and permutations of words and phrases are virtually infinite.

Even in use cases where all predictions can be generated in advance, it can be extremely wasteful to do so. Consider a recommender system designed to predict items for customers as an example. Generating a massive number of records periodically would not only be costly in terms of resource computation and usage, but it would also be time-consuming to generate millions of predictions. Why generate all predictions in advance if you can generate each prediction as needed?

This is where online predictions with batch features come in. Instead of precomputing all possible predictions before requests arrive, online predictions are served as soon as requests arrive. This is also known as on-demand prediction. Online serving can help prevent redundancy. For example, by generating predictions only for users who are visiting your website.

getinada-architecture-online-prediction

Very common batch features used for online serving are embeddings. In machine learning, embeddings are a way of representing data as vectors or low-dimensional numerical representations that capture the key features of the data. These numerical representations are often learned through neural networks and can be used to represent a variety of types of data, such as words, images, or even entire documents. By using embeddings, machine learning models can more easily process and understand complex data, leading to better performance in tasks such as classification, clustering and recommendation systems.

However, moving your models from offline to online presents some challenges, the first of which is inference latencySeveral studies show that website speed improvements can significantly affect business metrics such as conversion rates or average order value. Therefore, we need models that can generate predictions at a reasonable speed. There are several ways to keep your latency low, including selecting only the most informative features, pre-computing features, using simpler or smaller models, using more compute resources, and using model quantization. Another challenge is ensuring sufficient throughput and availability. Once you deploy your model, it must always be available with good uptime.

As you can see, online deployment can be one of the more complicated ways of deploying models.

Online predictions with streaming features

Static features also have limitations, as they make your models less responsive to changing environments. Switching to online predictions with streaming features allows you to use dynamic features to make more relevant predictions. For many problems, you need both streaming and batch features, which can be joined together in a single data pipeline and fed into your machine learning models.

Streaming features are especially useful in mission-centric contexts, where users visit your application with a pre-existing goal. Providing them with the right content or functionality in real-time is critical to keeping them engaged and preventing them from switching to competitors.

Let's say you're booking a hotel for your next vacation. Previously, you visited a specific country, and the website remembers this information, recommending hotels based on your previous travel destination. However, you want your recommendations updated based on the current context, considering any changes in your travel preferences or interests since your last trip. The recommendations may not be as relevant or useful if the website only relies on static features like your previous travel history. In this case, it would be beneficial to incorporate dynamic features that consider real-time data such as your current location, search history, and recent interactions with the website. By doing so, the website can provide more personalized and up-to-date recommendations that better reflect your current travel preferences.

getindata-streaming-batch-features

Nevertheless, incorporating streaming data adds significant complexity to your applications. It would help if you had appropriate real-time pipelines to utilize real-time predictions from streaming data fully. Technologies such as Apache Flink or Apache Kafka enable the computation of streaming features. However, stream processing poses unique challenges due to the unbounded amount of data that arrives at variable rates and speeds.

Moreover, online predictions are designed to make real-time decisions, meaning they can immediately impact the system or application they are supporting. This means that model monitoring for errors or negative feedback loops is critical to prevent issues from quickly escalating and causing further damage.

While real-time processing may seem attractive, it is only sometimes straightforward and can come with additional costs and complexities. For many common use cases, batch processing is an excellent approach, as it is highly efficient. In contrast, stream processing is fast because it allows you to process data as soon as it arrives. However, it is important to carefully consider the trade-offs before adopting true real-time streaming. Only do so after identifying a business use case that justifies the additional complexity and cost compared to batch processing.

How can you leverage Real-Time Machine Learning for Fraud Detection? A Case Study

Let's start with the definitions. What really is fraud detection?

Fraud detection is the process of identifying and preventing fraudulent activity within a system, organization or network. Fraud can take many forms, such as:

  • credit card fraud,
  • insurance fraud
  • identity theft,
  • phishing,
  • cybercrime.

Fraudulent activities can result in significant financial losses, reputation damage and legal consequences for individuals and organizations. Fraud detection involves the use of various methods and technologies to detect and prevent fraudulent activity in real-time, including rule-based systems and machine learning algorithms. The goal of fraud detection is to identify and prevent fraudulent activity before it can cause harm, and to help ensure the integrity and security of systems and organizations.

We will focus on credit card fraud as this is the most common form of fraud in the industry.

Credit card fraud is a type of financial fraud that involves the unauthorized use of a credit card to make purchases or cash withdrawal. Here's how it typically works, step by step:

  1. Obtaining the credit card details.
    The fraudster first obtains the credit card details, such as the card number, expiry date and security code. This can be done through various means, such as stealing the physical card, phishing scams or malware attacks on a point of sale system.
  2. Testing the card.
    The fraudster may then test the card to see if it works by making small purchases or online transactions to see if the card is still active and has not been reported as stolen.
  3. Making fraudulent purchases.
    Once the card is confirmed to be working, the fraudster may then use it to make larger purchases or cash withdrawals. They may use the card to buy goods or services online, over the phone or in person at a retail store or restaurant.
  4. Hiding the fraud.
    To hide the fraud, the fraudster may use various methods such as changing the billing address or shipping address, making multiple small transactions, or using different cards or accounts to make the purchases.

Examples of credit card fraud may include:

  • the card magnetic strip information is obtained through a breach when the card is used at an ATM in a shopping mall and then used from a geographically remote location within a short timeframe
  • during an online transaction, credit card information is stolen, and subsequently irregular transactions occur with varying amounts or at unusual times of the day after a certain period of time has passed
  • the credit card is physically stolen from a restaurant. A small transaction is initiated shortly after the theft to validate the card's usability, followed by a significantly larger transaction at a later time
  • an excessively large purchase has been made on a credit card
  • a transaction has been made from a location that is entirely different from the previous ones

As a result, customers will be charged for items they did not purchase. It's important to note that credit card fraud can have serious consequences for both the cardholder and the issuer. Credit card fraud can have severe consequences, including but not limited to:

  • financial loss for both the cardholder and the financial institution
  • loss of trust and confidence in the financial system by consumers
  • decline in credit scores resulting from undetected fraudulent activity over a prolonged period
  • legal repercussions for the cardholder, such as criminal charges and fines, as well as for the financial institution, which may face legal action from affected customers or regulatory authorities

If fraudulent activity is detected, the cardholder is typically notified, and the fraudulent transactions are reversed. The card issuer may also block the card and issue a new one to the cardholder. Protecting credit card details and reporting any suspicious activity as soon as possible is essential.

If fraudulent activity is detected, the cardholder is typically notified and the fraudulent transactions are reversed. The card issuer may also take steps to block the card and issue a new one to the cardholder. It's important to take steps to protect credit card details and report any suspicious activity as soon as possible.

Why real-time ML?

In today's world, transactions are processed almost instantly, providing customers with a seamless experience. However, this speed of processing also leaves banks and payment processors with a shorter time frame to identify and prevent fraud. In cases of online fraud, a thief may execute several transactions, resulting in significant financial losses. Therefore, it is essential that fraud detection occurs in near real-time. Delaying fraud detection until the next day may make it more challenging to recover the client's money.

Real-time fraud detection is critical because it allows businesses to respond quickly to real-time threats, preventing financial losses. Without real-time detection, fraud may only be detected after a financial loss has already occurred. Therefore, it is imperative to detect and react to real-time threats as they happen.

Fraud detection, much like cybersecurity, is an ongoing battle between attackers (fraudsters) and defenders (fraud detection systems). Fraudsters are constantly seeking new ways to profit from their malicious activities, while fraud detection systems are continuously challenged to stay one step ahead of them.

Machine learning algorithms can enable the dynamic detection of fraudulent transactions by analyzing large amounts of data to identify anomalous patterns associated with fraudulent activity. Incorporating streaming features into these models can enhance their responsiveness to changes. This means that financial institutions and cardholders can be immediately notified of suspicious transactions, allowing them to take swift action to prevent further fraudulent activity.

A total of $28.58 billion credit card fraud loss was suffered by credit card issuers, merchants, and consumers in 2020.

Real-time features can significantly enhance the performance of our model, which can have a major impact on credit card fraud prevention. Even a slight improvement in accuracy can translate into substantial savings, amounting to millions of dollars! In addition, features related to fraud detection require low latency processing from real-time data sources to ensure that the feature values remain fresh and up-to-date.

Here are some real-time features that we can consider to improve our machine learning models:

  • the number of transactions in the X hours
  • sudden spike in the number of transactions within a short period could be a sign of a coordinated attack or fraudulent activity (when a fraudster gains unauthorized access to the credit card, typically he will try to execute as many transactions as possible in the shortest amount of time; they hope to gain as much until the defenders will notice the abnormal behavior)
  • the number of transactions in the same place in the X hours
  • the occurrence of somebody making multiple purchases at the same store within a short period (tracking the number of transactions in that location over time can help us detect abnormal behavior)
  • the number of distinct geolocations in the last X hours
  • the number of unique geographic locations associated with a set of transactions (it helps us identify transactions that are impossible to execute for a human; for example, you can't buy from a shop in London and 15 minutes later in Tokyo)
  • time difference between the latest transaction
  • ratio of failed transactions in the last X days or the number of failed trades in a row (a higher ratio of failed transactions can also indicate a higher probability of a credit card being stolen)
  • requests to increase the credit card limit in the last hour
  • the fact of increasing credit card limit (often credit cards come with low limits, so fraudsters may attempt to increase the credit card limit on a stolen card to make larger purchases or transfer more funds)
  • the number of transactions in the specific location
  • a sudden spike in transactions in a specific location which can be a warning sign of fraudulent activity, as fraudsters may attempt to use a specific location to carry out their fraudulent transactions; this feature can be useful for identifying instances of organized crime or fraud rings, where a group of fraudsters may be coordinating their activity in a particular location or area
  • the fact whether the current transaction is significantly greater than the average of previous transactions

Architecture

This is an example of what a real-time fraud detection system could look like.

real-time fraud detection system getindata

It would work as follows:

  • when the user makes a transaction using their credit card, the transaction information is sent to the Transaction Backend
  • from there, the information is forwarded to both:
    the Fraud Detection System
    the Streaming Technology
  • the Fraud Detection System requests features from both the Online (Streaming) Data Source and the Offline (Batch) Data Source
  • streaming low-latency database contains real-time features
  • offline database contains static features like information about shop and user
  • after receiving the necessary features, the Fraud Detection System sends a request to the Machine Learning Model REST API, which uses a trained model to predict whether the transaction is fraudulent or not
  • the prediction is then returned to the Fraud Detection System, which communicates the results to the Transaction Backend

The software infrastructure must be designed to handle a high volume of transactions. The system must also be fault-tolerant and scalable to ensure it can handle increased traffic. Furthermore, it is essential to have proper monitoring and alert systems in place to identify and resolve any issues that may arise.

Summary

Overall, real-time detection is crucial for credit card fraud prevention, because it enables individuals and businesses to respond quickly to fraudulent activity, limit financial losses and prevent unauthorized access to funds. This can help prevent additional fraudulent activity, limit financial losses and provide a better chance of recovering lost funds.

Interested in ML and MLOps solutions? How to improve ML processes and scale project deliverability? Watch our MLOps demo and sign up for a free consultation.

batch processing
machine learning models
real-time ML
real-time fraud detection
real-time machine learning model
online prediction
8 August 2023

Want more? Check our articles

blogdzisssobszar roboczy 1 4
Tutorial

Deploying MLflow on the Google Cloud Platform using App Engine

MLOps platforms delivered by GetInData allow us to pick best of breed technologies to cover crucial functionalities. MLflow is one of the key…

Read more
writing flink jobs using springobszar roboczy 1 4

Writing Flink jobs using the Spring dependency injection framework

Introduction Almost two decades ago, the first version of Spring framework was released. During this time, Spring became the bedrock on which the…

Read more
getindata data democratization 2

Data Democratization: Power Your Organizations with Data Accessibility

In today's digital age, data reigns supreme as the lifeblood of organizations across industries. From enabling informed decision-making to driving…

Read more
getindata big data tech main 1
Big Data Event

A Review of the Presentations at the Big Data Technology Warsaw Summit 2022!

The 8th edition of the Big Data Tech Summit is already over, and we would like to thank all of the attendees for joining us this year. It was a real…

Read more
3

Data Journey with Michał Wróbel (RenoFi) - Doing more with less with a Modern Data Platform and ML at home

In this episode of the RadioData Podcast, Adam Kawa talks with Michał Wróbel about business use cases at RenoFi (​​a U.S.-based FinTech), the Modern…

Read more
runningkedroeverywhereobszar roboczy 1 4
Tutorial

Running Kedro… everywhere? Machine Learning Pipelines on Kubeflow, Vertex AI, Azure and Airflow

Building reliable machine learning pipelines puts a heavy burden on Data Scientists and Machine Learning engineers. It’s fairly easy to kick-off any…

Read more

Contact us

Interested in our solutions?
Contact us!

Together, we will select the best Big Data solutions for your organization and build a project that will have a real impact on your organization.


What did you find most impressive about GetInData?

They did a very good job in finding people that fitted in Acast both technically as well as culturally.
Type the form or send a e-mail: hello@getindata.com
The administrator of your personal data is GetInData Poland Sp. z o.o. with its registered seat in Warsaw (02-508), 39/20 Pulawska St. Your data is processed for the purpose of provision of electronic services in accordance with the Terms & Conditions. For more information on personal data processing and your rights please see Privacy Policy.

By submitting this form, you agree to our Terms & Conditions and Privacy Policy