Announcing Agent Commander - The First Integrated solution from Veeam + Securiti.ai enabling the scaling of safe AI agents

View
Veeam

The Funniest Evening at RSA with Hasan Minhaj

Hasan Minhaj Request ticket
View

GLBA Pretexting Rule – Guarding Customer Data Against Deceptive Practices

Contributors

Anas Baig

Product Marketing Manager at Securiti

Omer Imran Malik

Data Privacy Legal Manager, Securiti

FIP, CIPT, CIPM, CIPP/US

Published February 14, 2024

Listen to the content

The Gramm-leach-Bliley Act (GLBA) sits at the heart of the financial industry. The law regulates financial institutions, banks, insurance companies, and other entities dealing with non-public financial information (NPI). The law aims to protect customers' data from current and emerging threats and ensures its confidentiality, integrity, and availability.

However, with all the robust security measures and access policies in place, cybersecurity incidents, such as data breaches, tend to occur. Pretexting is one of the most common reasons behind such incidents, even when optimal security measures are implemented. This may be why the GLBA provides added provisions for such security threats.

This blog briefly explores pretexting, its common examples, and the relevant provisions outlined under the GLBA.

Understanding Pretexting

The cybersecurity sphere considers humans as the weakest link, a notion based on many compelling reasons. Take, for instance, the most prominent trait that renders humans susceptible to risks is curiosity. Threat actors understand this human trait very well and frequently exploit it through enticing traps. Pretexting is one such common yet highly effective trap that tops the list.

Pretexting is the illicit practice of accessing someone’s confidential or sensitive information using false pretenses. The cybercriminal creates a compelling situation that entices victims to divulge information they should not really expose. This practice is often experienced at the enterprise level, where scammers aim to target clients’ sensitive data, such as credit card numbers, account information, social security numbers, etc.

The threat actor attempts to request information from the company by impersonating someone with authority, such as a CEO, who can access sensitive information. One common example of pretexting is pretending to be an email provider. The cybercriminal may send a fraudulent email to a user, requesting them to verify their password or change it under the false pretext of a potential data breach. Consequently, the cybercriminal swiftly steals the password when the user clicks the provided link in the email and attempts to change the password.

Canadian University Phishing Scam

In this incident, the university fell victim to a fraudulent scheme amounting to $11.8 million. The perpetrators, posing as construction contractors, engaged in pretexting by creating deceptive websites and emails. Exploiting a lack of verification measures, the university failed to authenticate the legitimacy of the communication with the supposed contractors. Compounding the issue, the responsibility for communication with the contractor was delegated to three lower-tier employees. As a consequence of the phishing email, the university unwittingly altered the banking details of the construction contractors, resulting in a significant financial loss of nearly $11.8 million.

Techniques of Pretexting Scams

There are several techniques that cybercriminals may leverage to conduct pretexting attacks, such as:

Impersonation

Impersonation is amongst the most common types of pretexting attacks. In this attack, the attackers assume the identity of someone the victim trusts, such as a friend or a family member, or someone who is assumed to have access to sensitive information, such as an office employee, tech support staff, etc.

Phishing

Phishing is an ideal example of a pretexting attack. Phishing attacks are designed to deceive individuals into disclosing personal information or clicking on links that could result in malware infections. Unlike typical phishing tactics, social engineering pretexting frequently uses a targeted spear phishing approach, aiming to establish and prolong a connection over an extended duration.

Baiting

As the name suggests, baiting is luring or trapping an individual by making an attractive promise of a reward, such as a fraudulent refund or cash prize. When incorporated into a pretexting scheme, the scammer devises a scenario to instill confidence and a sense of security in the victim, encouraging them to take the bait. Usually, this type of attack helps the attacker steal credentials or sensitive information. This may include clicking a link or a URL in an email.

GLBA Pretexting Rule

GLBA Pretexting Rule has been designed to counter scammer attacks. GLBA prohibits the collection of information under false pretenses. The law stipulates that an individual must not obtain or try to obtain a customer’s NPI via false statements, fraudulent statements, or by falsely assuming the identity of a representative, an employee, or a customer. The law further prohibits individuals from intentionally forging or counterfeiting any documents to obtain customer information. The law goes on to penalize individuals with a fine, imprisonment, or both if the individual is found to be knowingly and intentionally attempting to violate the Pretexting Rule.

Best Practices to Prevent Pretexting Risks

The law provides no recommendations or guidelines for preventing or overcoming pretexting threats. However, there are certain best practices that financial institutions may consider to prevent such attacks and protect the integrity, confidentiality, availability, and accessibility of customers’ data. These practices may include;

Staff Training

As pretexting manipulates individuals into compromising their own security, providing training to employees on recognizing and responding to pretexting scams can contribute to safeguarding an organization. Employee training should be a priority for every organization managing sensitive information, such as financial data. Employee training should encompass protocols for managing sensitive information, organizational policies, and procedures. This training should also incorporate guidelines regarding access to information, aiding employees in distinguishing between authorized and unauthorized individuals.

Robust Authentication and Authorization Process

It is crucial for organizations to establish and implement a robust authentication and authorization process to ensure that only verified individuals can access sensitive data or accounts. Moreover, the organization must create and implement strict access policies and controls to ensure that authorized individuals can access customer data. It is recommended to strive for role-based access control (RBAC) policy and minimize access to only what is required.

Domain-based Message Authentication, Reporting, and Conformance (DMARC)

DMARC is an email authentication protocol designed to counteract spoofing by verifying whether an email originates from the claimed domain. In cases of spoofing email, the system can automatically redirect the email to a spam folder or delete it. This practice can be implemented to deter pretexting scams.

Streamline GLBA Compliance with Securiti PrivacyOps

Securiti PrivacyOps, an extension of the Data Command Center, enables financial institutions to streamline their GLBA compliance requirements with unified intelligence around data and automated controls.

Schedule a demo to see PrivacyOps in action.

Analyze this article with AI

Prompts open in third-party AI tools.
Join Our Newsletter

Get all the latest information, law updates and more delivered to your inbox



More Stories that May Interest You
Videos
View More
Rehan Jalil, Veeam on Agent Commander : theCUBE + NYSE Wired: Cyber Security Leaders
Following Veeam’s acquisition of Securiti, the launch of Agent Commander marks an important step toward helping enterprises adopt AI agents with greater confidence. In...
View More
Mitigating OWASP Top 10 for LLM Applications 2025
Generative AI (GenAI) has transformed how enterprises operate, scale, and grow. There’s an AI application for every purpose, from increasing employee productivity to streamlining...
View More
Top 6 DSPM Use Cases
With the advent of Generative AI (GenAI), data has become more dynamic. New data is generated faster than ever, transmitted to various systems, applications,...
View More
Colorado Privacy Act (CPA)
What is the Colorado Privacy Act? The CPA is a comprehensive privacy law signed on July 7, 2021. It established new standards for personal...
View More
Securiti for Copilot in SaaS
Accelerate Copilot Adoption Securely & Confidently Organizations are eager to adopt Microsoft 365 Copilot for increased productivity and efficiency. However, security concerns like data...
View More
Top 10 Considerations for Safely Using Unstructured Data with GenAI
A staggering 90% of an organization's data is unstructured. This data is rapidly being used to fuel GenAI applications like chatbots and AI search....
View More
Gencore AI: Building Safe, Enterprise-grade AI Systems in Minutes
As enterprises adopt generative AI, data and AI teams face numerous hurdles: securely connecting unstructured and structured data sources, maintaining proper controls and governance,...
View More
Navigating CPRA: Key Insights for Businesses
What is CPRA? The California Privacy Rights Act (CPRA) is California's state legislation aimed at protecting residents' digital privacy. It became effective on January...
View More
Navigating the Shift: Transitioning to PCI DSS v4.0
What is PCI DSS? PCI DSS (Payment Card Industry Data Security Standard) is a set of security standards to ensure safe processing, storage, and...
View More
Securing Data+AI : Playbook for Trust, Risk, and Security Management (TRiSM)
AI's growing security risks have 48% of global CISOs alarmed. Join this keynote to learn about a practical playbook for enabling AI Trust, Risk,...

Spotlight Talks

Spotlight 50:52
From Data to Deployment: Safeguarding Enterprise AI with Security and Governance
Watch Now View
Spotlight 11:29
Not Hype — Dye & Durham’s Analytics Head Shows What AI at Work Really Looks Like
Not Hype — Dye & Durham’s Analytics Head Shows What AI at Work Really Looks Like
Watch Now View
Spotlight 11:18
Rewiring Real Estate Finance — How Walker & Dunlop Is Giving Its $135B Portfolio a Data-First Refresh
Watch Now View
Spotlight 13:38
Accelerating Miracles — How Sanofi is Embedding AI to Significantly Reduce Drug Development Timelines
Sanofi Thumbnail
Watch Now View
Spotlight 10:35
There’s Been a Material Shift in the Data Center of Gravity
Watch Now View
Spotlight 14:21
AI Governance Is Much More than Technology Risk Mitigation
AI Governance Is Much More than Technology Risk Mitigation
Watch Now View
Spotlight 12:!3
You Can’t Build Pipelines, Warehouses, or AI Platforms Without Business Knowledge
Watch Now View
Spotlight 47:42
Cybersecurity – Where Leaders are Buying, Building, and Partnering
Rehan Jalil
Watch Now View
Spotlight 27:29
Building Safe AI with Databricks and Gencore
Rehan Jalil
Watch Now View
Spotlight 46:02
Building Safe Enterprise AI: A Practical Roadmap
Watch Now View
Latest
View More
Introducing Agent Commander
The promise of AI Agents is staggering— intelligent systems that make decisions, use tools, automate complex workflows act as force multipliers for every knowledge...
Risk Silos: The Biggest AI Problem Boards Aren’t Talking About View More
Risk Silos: The Biggest AI Problem Boards Aren’t Talking About
Boards are tuned in to the AI conversation, but there’s a blind spot many organizations still haven’t named: risk silos. Everyone agrees AI governance...
Largest Fine In CCPA History_ What The Latest CCPA Enforcement Action Teaches Businesses View More
Largest Fine In CCPA History: What The Latest CCPA Enforcement Action Teaches Businesses
Businesses can take some vital lessons from the recent biggest enforcement action in CCPA history. Securiti’s blog covers all the important details to know.
View More
AI & HIPAA: What It Means and How to Automate Compliance
Explore how the Health Insurance Portability and Accountability Act (HIPAA) applies to Artificial Intelligence (AI) in securing Protected Health Information (PHI). Learn how to...
California’s Delete Request and Opt-out Platform (DROP) and the Delete Act View More
California’s Delete Request and Opt-out Platform (DROP) and the Delete Act
Understand California’s DROP platform and the Delete Act, including compliance timelines, the 45-day cycle, broker obligations, and how to operationalize compliance.
Building A Secure AI Foundation For Financial Services View More
Building A Secure AI Foundation For Financial Services
Access the whitepaper and discover how financial institutions eliminate Shadow AI, enforce real-time AI policies, and secure sensitive data with a unified DataAI control...
Emerging AI Security Trends For 2026 View More
Emerging AI Security Trends For 2026
Securiti’s latest infographic provides security leaders with a walkthrough of all the emerging AI security trends for 2026 to help them assess and plan...
Safe AI, Accelerated: View More
Safe AI, Accelerated: Securing Data & AI Across the Lifecycle
Securiti’s latest infographic dives into the issue organizations face when scaling their AI projects safely, and how best they can address those challenges.
View More
Take the Data Risk Out of AI
Learn how to prepare enterprise data for safe Gemini Enterprise adoption with upstream governance, sensitive data discovery, and pre-index policy controls.
View More
Navigating HITRUST: A Guide to Certification
Securiti's eBook is a practical guide to HITRUST certification, covering everything from choosing i1 vs r2 and scope systems to managing CAPs & planning...
What's
New