Brought to you by:
Enterprise Strategy Group  |  Getting to the Bigger Truth™

ESG SHOWCASE

Securing Applications from Sophisticated Bot Attacks with HUMAN

By John Grady, Senior Analyst
APRIL 2021

ABSTRACT

Fraudulent activity caused by non-human, bot-based web traffic is impacting an increasing number of organizations. However, the sophisticated nature of these bots has made detection more difficult. Existing application security solutions have trouble detecting sophisticated bot activity specifically because it is designed to mimic human actions and compromise user accounts through legitimate application functions. Dedicated, multilayered bot detection and mitigation solutions have become a prerequisite to ensure web transactions are originating from a human, and not a sophisticated bot or malware infected device. HUMAN has spent nearly ten years protecting enterprises and the largest internet platforms from fraud caused by sophisticated bots, and today, HUMAN verifies the humanity of more than 10 trillion interactions per week. Its Application Integrity offering expands that protection to online enterprises to protect their websites and applications from sophisticated bots and automated attacks, including account takeover, new account fraud, payment fraud, and content and experience abuse.

Sophisticated Bots Represent an Increasingly Disruptive, But Often Overlooked Threat Vector

Most businesses today rely on e-commerce capabilities, customer-facing web applications, and open APIs that are susceptible to abuse by sophisticated bots. Yet while many organizations know they have experienced these types of attacks, the sophistication modern bots exhibit by mimicking human behavior and executing legitimate transactions that become malicious only by their scale or outcome can make detection difficult (see Figure 1).1 Some of the most common fraudulent and abusive bot-based attacks include:

ACCOUNT TAKEOVER (ATO)

Nearly one-third (31%) of ESG research respondents have experienced ATO attacks from sophisticated bots. In an ATO attack, existing user accounts are compromised and exploited by cyber-criminals, typically through credential stuffing or credential cracking, activities which can run at a high scale through sophisticated bots.

NEW ACCOUNT FRAUD

Almost half (45%) of respondents cite automated account creation by sophisticated bots as an attack they have seen. In this case, fraudulent accounts created by sophisticated bots can be used for social media disinformation, phony product reviews, and other reputation-based attacks, or more direct financial attacks such as inventory holding and spoofing, payment and wire fraud, and money laundering.

SENSITIVE DATA THEFT

One-third (33%) of respondents have experienced sensitive content scraping from their websites. While not all web-crawling comes from bots controlled by hackers, malicious scraping may target the data layer to steal pricing, intellectual property, and customer information.

Figure 1. Prevalence of Bot Attacks

Has your organization been impacted by sophisticated bots over the past 12 months? (Percent of respondents, N=425)

Source: Enterprise Strategy Group

The initial account takeover or creation is only the first stage of the attack and is typically followed by fraudulent purchases, fund extraction, or other financially impactful actions. Alternatively, hackers can sell the validated compromised account information for a premium on the black market. In fact, account access-as-a-service is now commonly seen on the dark web with criminals offering managed access to legitimate services, providing buyers with the benefits of the compromised accounts without the risk or effort involved in the actual takeover.

The Availability of Stolen Credentials, Improvements in Automation, and Insecure Devices Have Created a Perfect Storm for Sophisticated Bot Fraud

Bots are not a new threat vector, so why has this become such a problem? To begin with, the universe of potential targets is massive and continues to expand with the acceleration of digital transformation. While e-commerce or financial entities are typically top of mind when thinking about the impact of bot activity, any organization with a web presence (such as healthcare, insurance, education, technology, travel, gaming, and entertainment) is a potential mark for a bot operator. Additionally, bot technology and the surrounding ecosystem has improved the success rate and economics of attacks:

Sophisticated bots can pass for legitimate humans – This was not always the case: the first and second generation of bots were fairly easy to detect due to their inability to store cookies or execute JavaScript, and the presence of easily detectable characteristics signifying the presence of automation. However, the third and fourth generation bots in use today are exceptionally hard to identify through browser and device analysis alone. They look and act like humans and originate from legitimate browsers. Additionally, the utilization of artificial intelligence (AI) by threat actors is on the rise and further complicates detection.
• Unsecure consumer devices contribute to bot sophistication and scale – Part of the reason bots are able to so closely mimic human activity and originate from seemingly innocuous residential IP addresses is because they exploit compromised consumer devices. Criminals have access to an untold number of malware-infected devices across the globe. These can be used by bots to track, study, and incorporate real-world human activity, such as non-linear mouse movements. Also, by routing bot traffic through the residential IP addresses of these devices or compromised IoT devices, IP reputation-based detection methods lose their effectiveness. It also provides cheap and easy access to bandwidth to massively scale attacks.
• Stolen credentials are numerous and cheap – Tens of billions of credentials from successful attacks are available on the dark web, with as many 7.9 billion records exposed in the first nine months of 2019 alone.2 Attackers can cheaply purchase large quantities of these credentials and, with their sophisticated bots, launch stuffing or cracking attacks leading to ATOs, knowing there is a good chance MFA is not in use to secure authentication. In fact, 31% of ESG research respondents have experienced a credential cracking or similar brute force attack in the last 12 months.
Where Legacy Solutions Fall Short in Sophisticated Bot Mitigation

While awareness of the threat posed by sophisticated bots is growing, existing security tools do not adequately address the issue. Historically, simple bots were focused on web scraping or in some cases would launch a high volume of fraudulent login attempts or account-based attacks. These attacks were basic and straightforward to detect and mitigate via traditional application security tools or manual intervention because of the noticeable volume or source IP addresses used. However, as bots have grown more sophisticated, these defenses have become less effective. Application security has historically included both testing and runtime protection solutions. Testing includes both static and dynamic application security testing (SAST and DAST), while runtime protection includes web application firewalls (WAFs) and runtime application self-protection (RASP). Testing solutions scan the code or production application itself to discover potentially exploitable vulnerabilities, misconfigurations, or authentication issues. Because bots exploit legitimate application operations as opposed to vulnerabilities or misconfigurations, these solutions do not address bot-based threats.

86% of ESG research respondents agree that most bots are capable of bypassing the simple protection offered as a feature in other products..."
WAF and RASP solutions are more applicable to the simple bot threat, and some vendors have attempted to add functionality to address more sophisticated bots. However, these solutions are ultimately built and deployed to solve a problem other than bots and lack the significant resources needed to remain ahead of attackers and recognize and stop sophisticated bot attacks. Some of the drawbacks of WAFs related to mitigating sophisticated bot attacks include:
• Limited rulesets – WAFs are often deployed simply to maintain PCI compliance, which may mean they are run with a limited ruleset to prevent false positives from impacting user availability, thus decreasing bot detection capabilities.
• Vulnerability focus – WAFs are typically focused on protecting against known software vulnerabilities (such as SQL injections and cross-site scripting). Bot-based attacks do not exploit these vulnerabilities, rendering this focus ineffective in bot mitigation.
• Signature dependency – WAFs have historically relied on signatures for detection. This is changing with third generation solutions but remains a component of many WAFs and is an insufficient method for detecting sophisticated bot attacks.
• IP reputation-centric – WAFs also rely on IP reputation to identify traffic from known-bad actors. Many sophisticated bots cycle through residential IP addresses, limiting the effectiveness of a reputational defense.

Yet despite the fact that 86% of ESG research respondents agree that most bots are capable of bypassing the simple protection offered as a feature in other products, three quarters (76%) indicate they use the features in other application security tools or platforms as their core approach to bot management.

Protecting Applications from Sophisticated Bots Requires Multilayered Solutions

Protecting applications against fraudulent activity is as important as maintaining the integrity of the application itself and a vastly different function. Rather than focusing on code, solutions must weigh intent as it relates to identity and potentially abusive activities. Brand perception and customer trust are at stake. To this end, a third subcategory of application protection solutions focused on anti-fraud extends beyond vulnerability and code-based exploit mitigation and protects against the malicious use of legitimate application functions (see Figure 2). Sophisticated bot and automated attack detection to prevent fraudulent activity is a key component of this subcategory.

Figure 2. Application Protection Subcategories

Because sophisticated bots are specifically mimicking the behavior of normal users, multiple layers of detection are required to ensure efficacy while maintaining a low level of false positives. Rather than focusing solely on the action the entity is taking within the application, detection techniques must consider the surrounding context to accurately detect malicious activity. To this end, the following detection mechanisms, used in conjunction with one another, represent a layered approach to detecting sophisticated bots by determining the intent of the entity behind the request.

Indicators of compromise – As discussed, today’s sophisticated bots emulate browser activity or automate within legitimate browsers originating from benign residential IP addresses, making detection difficult. JavaScript challenges and browser fingerprinting can still identify bots on their own. However, collecting and analyzing a broad set of information about the user device sending the request, the network it is originating from, and the software and applications running on the device helps detect more sophisticated bots with greater accuracy.
Analytics and continuous iteration – Bots continuously adjust and adapt to detection techniques, making it necessary to cycle and update tests and markers to stay ahead of attackers. Further, incorporating machine learning and analytics to correlate potentially disparate pieces of information can assist in detection even when individual indicators of compromise are not present.
Threat intelligence – riving continuous iteration requires a global mechanism to collect up-to-date bot-driven fraud techniques and contextual reputation data. As discussed, IP reputation alone does not work. However, considering patterns, determining the presence of proxies, and correlating across a broad ecosystem of customers and transactions can provide additional context into decision making.

This approach is resonating with users. In fact, despite so many organizations currently using other application security tools for bot management purposes, a strong majority (60%) feels that specialized solutions are most effective in preventing impacts from sophisticated bot attacks.

Enter HUMAN

Founded in 2012, HUMAN (formerly White Ops) protects both global enterprises and the largest internet platforms from bot attacks. The vendor has historically focused on combating advertising and marketing fraud and has expanded its capabilities to help organizations defend against the broad range of sophisticated attacks generated by bots. Its Human Verification Engine leverages HUMAN’s multi-layered detection technology, global visibility, and threat intelligence resources to verify the humanity of ten trillion interactions per week across digital advertising, performance marketing, and web and mobile applications.

Introducing Application Integrity

Powered by HUMAN’s Human Verification Engine, Application Integrity leverages HUMAN’s broad detection capabilities to protect websites and applications from sophisticated bots and automated attacks. Application Integrity utilizes a multi-layered decision engine incorporating more than 350 algorithms to detect signs of automation, remote control, and manipulated human activity. First, signal collection provides a granular understanding of individual transactions through JavaScript payload or software development kit (SDK) for mobile applications. Customers can also send signals to Application Integrity such as session IDs, hashed user IDs, referrers, and timestamps for complete feedback loops into user account activity.

Next, marker analysis runs the requesting device through hundreds of tests to detect automation. The decisioning engine is designed so that the failure of even one test causes the device to be considered a bot with a high degree of confidence, avoiding false positives. Finally, new tests are continually developed, and existing tests are adjusted, making it even harder for fraudsters to evade detection. The decision engine is supported by machine learning and global threat intelligence capabilities to detect bots even with limited technical evidence. Machine learning capabilities are driven by the more than ten trillion weekly transactions HUMAN has visibility into. HUMAN threat intelligence capabilities enable it to uncover emerging bot-based threats and attribute them to specific botnet operators, campaigns, and threat actors.

The Bigger Truth

In the hierarchy of threats, sophisticated bot activity has flown under the radar when compared with attacks such as advanced malware, distributed denial of service, and, more recently, ransomware. However, with the negative impacts resulting from these types of attacks ranging from poor customer experience to lost revenue and shareholder value, the issue deserves more attention. It appears that organizations are beginning to understand the significance of this issue, as protecting against sophisticated bot attacks is now a top 3 priority for more than 60% of security teams. Because of the availability of stolen credentials, compromised machines, and sophisticated bot technology, we appear to be at an inflection point. Security and risk leaders are starting to understand their existing application security tools were not built to detect sophisticated bot activity. To round out application security capabilities, security leaders should consider not just testing and runtime protection, but anti-bot solutions as well. With a history of protecting organizations from bot-based fraud, HUMAN is uniquely positioned to continue expanding its capabilities into the application security space.

Know Who's Real

This ESG Showcase was commissioned by HUMAN (formerly White Ops) and is distributed under license from ESG.

ESG eBook commissioned by HUMAN (formerly White Ops), 2021 Bot Management Trends: Harmful Attacks Drive Interest in Specialized Solutions. All ESG research references and charts in this showcase have been taken from this eBook, unless otherwise noted.

Source: Risk Based Security, Q3 2019 Data Breach QuickView Report, Nov 2019.

Source: ESG Master Survey Results, Transitioning Network Security Controls to the Cloud, Jul 2020.

Source: ESG Master Survey Results, Network Security Trends, Mar 2020.

All trademark names are property of their respective companies. Information contained in this publication has been obtained by sources The Enterprise Strategy Group (ESG) considers to be reliable but is not warranted by ESG. This publication may contain opinions of ESG, which are subject to change from time to time. This publication is copyrighted by The Enterprise Strategy Group, Inc. Any reproduction or redistribution of this publication, in whole or in part, whether in hard-copy format, electronically, or otherwise to persons not authorized to receive it, without the express consent of The Enterprise Strategy Group, Inc., is in violation of U.S. copyright law and will be subject to an action for civil damages and, if applicable, criminal prosecution. Should you have any questions, please contact ESG Client Relations at 508.482.0188.

Enterprise Strategy Group | Getting to the Bigger Truth™

Enterprise Strategy Group is an IT analyst, research, validation, and strategy firm that provides market intelligence and actionable insight to the global IT community.