Back to the blog
September 24, 2025

AI discovery methods compared: A 2025 guide to finding and managing shadow AI

With AI making it’s way into virtually every SaaS application, AI discovery extends far beyond chat prompts and purpose-built AI tools.

The generative AI landscape is changing fast. In just the last year, Nudge Security has discovered over 1,000 unique new AI tools hitting the market, not to mention the MCPs and AI capabilities being added to virtually every other SaaS tool used by your workforce. Given this pervasiveness, it’s critical to think about the full SaaS ecosystem when evaluating solutions to help you discover workforce AI use and mitigate risks.

But with a growing AI security market and a maze of vendor claims, how do you separate real AI visibility and control from smoke and mirrors? Let’s cut through the noise and break down the most common AI discovery methods, so you can find the right fit for your organization.

What is AI discovery?

AI discovery is the process of identifying and cataloging all AI apps, accounts, integrations, user activities, and other dependencies that could expose your corporate data to third-party AI providers and beyond. An AI asset inventory includes AI apps that are procured and sanctioned by IT as well as “shadow AI”—the long tail of apps and embedded AI features that your employees experiment with and adopt without going through a formal approval process.

Not unlike the asset discovery and inventory systems developed for enterprise networks, the purpose of AI discovery is to create a comprehensive system of record of all AI assets introduced across your organization. But, instead of assets like servers, workstations, and installed software, your AI asset inventory captures third-party AI apps, user accounts, authentication and entitlement data, integrations across apps, MCPs, and AI supply chain dependencies.

Effective AI discovery can help you answer important questions like:

  • Who is using AI across my organization?
  • Have new (unapproved) AI tools been introduced outside of our governance process?
  • Are AI tools integrated with apps that store our business critical data, and what level of access do these tools have?
  • Are AI-enabled features in our other workplace technology solutions making our data accessible to LLMs?
  • Is AI used in the supply chain of our other SaaS providers, and if so, what assurances have they provided on their data privacy practices?

Step one of defining your AI discovery approach is to define the scope of what you need to discover. A common mistake is to focus only on discovering AI prompt activity in chatbots like ChatGPT, which is merely the tip of the iceberg given that AI is now making it’s way into virtually every other SaaS tool used by your workforce. This guide aims to provide a more comprehensive view of how AI tools and providers could gain access to your corporate data, so you can prioritize your efforts accordingly.

Note: In the interest of avoiding scope creep, this guide focuses specifically on workforce AI use. In other words, the AI tools and embedded AI that can gain access to your sensitive data based on use by your employees. AI model security and agentic AI security are related topics, but warrant their own separate deep dive.

What should AI discovery include?

When it comes to workforce AI use, here are the essential categories of AI assets you should consider for your discovery efforts:

1. Purpose-built AI apps

These are apps like ChatGPT, Perplexity, Claude, and others that offer a user interface for your workforce to interact with the LLM in order to answer questions, analyze data, get help writing, conduct research, generate images, and more.

Ideally, you want to discover all user accounts for all apps in this category. Not an easy feat given that the number of purpose-built AI tools is increasing exponentially.

*Tip: Look for solutions that can pattern match to recognize AI tools regardless of whether you’ve heard of them or not. Any static or vendor-managed list of AI tools will quickly become outdated.

2. Integrations that grant AI tools access to other business apps and data (i.e. OAuth grants)

Many workforce AI tools offer easy connections to others platforms via OAuth grants or native marketplace apps. The ease of approving an OAuth grant or API key can entice users to hand over more access than they might realize—a well-meaning employee can inadvertently give AI tools access to, say, their entire corporate Google Drive, calendar, or email without realizing it.

A recent example of this is AI notetaker apps that prompt users to grant access to calendars and contacts, while defaulting to “join every meeting.” Suddenly, unapproved AI notetakers are popping up in meetings where sensitive topics are discussed, or where attendees have not consented to being recorded.

*Tip: Look for solutions that can give you a robust inventory of app-to-app integrations through OAuth grants, APIs, and other methods, including details on scopes and potential risks, ideally with the ability to directly revoke risky access.

3. MCP server integrations

Recently, we’ve seen a massive shift in how corporate data is accessed by AI tools following Anthropic’s release of the open-source model context protocol (MCP). Major SaaS providers like Github, Atlassian, Zapier, and Hubspot have followed suit, announcing their own MCP servers and native AI integrations through their own marketplaces.

Instead of relying on one-off prompts and individual file uploads, these MCP servers radically expand data access by allowing LLMs to directly query SaaS apps through backend APIs. While this expanded data set improves the quality and value of AI outputs by grounding them with your corporate data, it introduces significant new data security risks.

MCP and other direct connections occur app to app, directly from the backend of providers like OpenAI to the data stores like Salesforce that have been integrated through OAuth grants or APIs. And, they typically grant the AI tool the same data permissions as the user who created the connection, which can result in much broader data access than what the user intended, or what your data security policies allow.

So, in addition to the OAuth inventory mentioned above, discovering and governing API connections to AI tools is also critical in this new MCP-powered phase of AI use.

*Tip: SaaS Security Posture Management (SSPM) tools can be helpful here as they typically inventory the integrations that have been enabled within the apps you have connected to the SSPM, so you don’t have to manually review each app-to-app integration.

4. SaaS apps with embedded AI

The next category that is often overlooked, but is actually growing the fastest is SaaS apps that have embedded AI functionality within them. SaaS providers have been racing to launch AI-enabled features within their products and users have been equally eager to experiment with these new capabilities. However, enabling these embedded AI features may result in data access for the underlying LLM beyond what you want to allow.

*Tip: Some SaaS and AI governance solutions can surface and summarize the data training policies of your SaaS and AI vendors, including whether your data is used for model training, available opt-out options, retention periods, and other relevant information, saving you the time and effort of manually reviewing lengthy legal documentation.

5. AI in the supply chain of your other app providers

Another category that is often overlooked is AI in the supply chain of other SaaS app providers. An example of this would be companies who use AI tools to streamline customer support operations. Let’s say you are a customer of Acme SaaS tools, and they use a third-party AI provider as part of their customer support process. If sensitive data from your environment is shared (say, tokens or API keys) as part of a support case, that data could now be accessible by the underlying LLM.

*Tip: Ask your SaaS providers key questions about their third-party AI use, such as “Do you process our data through a third-party AI provider? If so, what security controls do you have in place for protecting our data privacy? What options do we have to limit what access could be accessible?”

6. AI user activities like chat prompts and file uploads

Last but not least, you’ll want to consider what data or files users might be sharing with AI tools via chat prompts or file uploads. Typically, detecting these activities requires a browser-, endpoint- or network-based control that can recognize file-upload behavior or sensitive data being uploaded (like API keys or credit card numbers) and take action to alert the user of potential data security risks, notify the IT security team of this activity, or possibly intercept the prompt and obfuscate the sensitive data before it reaches the LLM.

*Tip: Network-based controls are only useful to the extent that your employees work from within a “walled garden” visible to those controls. Browser-based controls, or a layered approach (which we’ll cover below), tend to be more effective given the highly distributed nature of how and where modern work gets done.

Do you need a new tool for AI discovery?

In most cases, this answer will be an emphatic “yes,” but there are some aspects of AI discovery that you can get started with using tools you probably already own. Here’s how you can detect AI activity manually:

Expense reports and financial systems

If someone is paying for an AI tool, it will likely show up in your expense tracking systems at some point. While this data will show you paid tools, it will obviously miss free trials or free tier usage. Additionally, it might show you who the billing contact is, but you will still have a lot of digging to do to identify all the users of these paid tools.

Network logs

Some customers set up rules within their network monitoring tools to look for signs of AI usage based on a specific list of known AI domains, or heuristics like domains ending in “.ai." This can provide better granularity with regard to the actual users of an AI tool and how frequently they are accessing them, but the downside is you will have to either know what AI tools to look for, or be ready for some manual effort to separate the actual AI tools from the false positives. And, if your workforce is largely remote, this method provides even more limited value.

IdP OAuth grants

If your employees commonly use the “sign up with Google” or “sign up with Microsoft” options when they create accounts for new apps, you’ll be able to see these by reviewing the OAuth grants that have been established via your IdP. This is helpful, but of course only works if the user signed up with their work email, and opted for this option rather than a username and password.

This method will also show you where users may have connected AI apps to the suite of tools within the IdP, such as their calendar, email, contacts, or file shares.

As with the network log problem, you will still need to develop a system for determining which tools are AI tools. And, OAuth grant reviews can be time consuming and tedious without some automated method to help you find grants for AI tools, understand the scopes that have been grated, and assess the legitimacy of the grant.

SSO activity

Obviously, your SSO platform can only show you accounts and activity for the AI apps that have actually been onboarded into SSO, but this can be useful if you are trying to understand which AI tools are used most often and by which users and track the progress of AI adoption efforts. But, when it comes to uncovering unsanctioned AI use, this method will not be as helpful.

Individual app integrations

For your most critical apps (i.e. those in scope for your compliance efforts), you could go into the integration settings for each app to review OAuth grants and API connections that have been enabled for signs of connections with AI tools. Again, this would involve manual effort and regular audits, which limits the scalability and timeliness of the data.

If the SaaS tools (and your subscription tier for the tool) include logging of activities like enabling integrations, you could forward these events to your SIEM, SOAR, or other security monitoring tools to alert you when new integrations are enabled.

*The DIY summary: The obvious benefit of using the AI discovery methods outlined above is avoiding the expense of adding another tool to your tech stack, but the data you can gather is limited, likely to become quickly outdated, and will require significant manual effort. Unless you operate in an environment with very strict controls over incoming and outgoing connections, you should likely consider a more complete and scalable solution.

How to discover workforce AI use

Given the challenges outlined above, it’s no surprise that the marketplace of security solutions is evolving rapidly to address the emerging risks posed by workforce AI use. Given the myriad of vendors and claims, it’s important to understand the types of tools that are emerging to help with this challenge and the nuances of what they can and can’t do.

At a high level, there are essentially two categories of tools to consider:

  • SaaS security solutions: these platforms can discover use of shadow SaaS, and have expanded their capabilities to discover and categorize AI use, along with providing other AI security and governance capabilities.
  • AI security point solutions: a new and rapidly expanding list of tools designed specifically for governing workforce AI use and protecting sensitive data.

Let’s start with SaaS security solutions as there are differences in how these tools actually discover AI use that are important to understand. SaaS security solutions rely on one (or more) of the following methods to discover SaaS and AI use:

  • Network-based monitoring and/or endpoint agents
  • Browser-based discovery
  • Email-based discovery
  • API connections with SaaS apps
  • Integration with HRIS, financial or CASB systems

Here’s the rundown on what you need to know about how each discovery method works.

Network-based monitoring and endpoint agents (for example, CASBs)

These approaches rely on capturing user traffic—either via desktop agents installed on individual devices or network-level monitoring through firewalls, VPNs, or Cloud Access Security Brokers (CASBs). Similar to the network log monitoring option in the “DIY” section, this method detects and logs access to domains, providing visibility into what apps employees are using based on their web traffic.

Network-based monitoring and CASB solutions can theoretically provide broad visibility into AI tool usage, but in practice they face significant challenges in modern work environments. These methods struggle with decentralized workforces, BYOD policies, and the technical limitations of detecting modern AI applications with dynamic domains and encrypted traffic. While they may have their place in specific environments with tight network controls, for agile organizations with remote workers, they're often not the right solution.

How it works:

Traffic inspection via network-level monitoring or desktop agents that flags traffic to the domains of known AI providers. Forward-looking only.

Pros:

  • Monitors cloud service access
  • Applies security policies

Cons:

  • Limited to corporate networks
  • Challenging for remote work
  • Resource-intensive
  • Requires endpoint agents
  • No historical discovery

Questions to ask of vendors:

  • Do you offer built-in rules to identify AI use? If so, how frequently are those rules updated?
  • What is the average deployment time for your solution? Average time to value?
  • Does your solution offer both inline and out-of-band (API-based) discovery and security monitoring?
  • Does your solution detect app-to-app integrations, including MCP connections?

Browser-based AI discovery

Browser-based AI discovery monitors app usage directly through a lightweight browser extension deployed to corporate devices. Extensions installed in corporate browsers can detect AI web visits, account signups, login activity, authentication methods, password strength, app usage patterns, file sharing, prompt content, and other rich behavioral and risk insights.

While browser extensions can offer robust visibility into AI activity after installation, their only potential window into historical AI activity comes from browser history data, a low-fidelity source for AI discovery. Additionally, some browser extensions identify AI activity based on domain visits rather than actual login data, offering extremely limited security-relevant information, while others only offer support for a specific set of AI domains, limiting their utility for AI discovery and other essential use cases.

How it works:

Monitors for browser activity happening between users and the domains of AI tools.

Pros:

  • Granular user activity data
  • Real-time interventions
  • Usage pattern insights

Cons:

  • Misses mobile/personal device usage
  • Limited historical data
  • Depends on installation
  • May have limited domain support

Questions to ask of vendors:

  • Can you differentiate domain visits from login data?
  • What metadata does the browser extension collect?
  • How are you identifying activity related to AI tools vs. other SaaS apps?

Email-based AI discovery

This approach analyzes corporate email communications from SaaS and AI providers for evidence of AI activities—things like welcome emails, password resets, billing notifications, MFA prompts, and security alerts. This discovery method works via read-only API connections to your IdP provider (like Google Workspace or Microsoft 365) in order to scan for emails related to AI app usage, identifying both the sanctioned and unsanctioned tools users have signed up for. And, advanced algorithms pattern match to recognize AI tools versus other types of SaaS apps.

The clear advantage of email-based discovery and automated classification is the ability to discover AI apps without needing to tell the discovery engine what to look for. Even better, this discovery method can uncover past AI use based on the email history of your users. So, instead of having to wait for new activity to occur before you can take action, solutions leveraging this discovery method can provide a full inventory of AI apps and accounts within minutes to hours of enabling the email integration.

The biggest limitation? Email discovery can’t detect AI activity tied to personal email accounts. If someone uses their Gmail or iCloud address to sign up for a tool or doesn't sign up at all, that won’t appear in a corporate inbox—and won’t be detected unless paired with another method like a browser extension.

Still, email-based discovery is one of the fastest and most scalable ways to uncover the long tail of shadow AI across your organization.

How it works:

Connects to your email provider and analyzes historical and future email activity for signs of AI use and captures rich context related to this activity.

Pros:

  • Broad discovery coverage
  • Finds unknown applications
  • Historical insight
  • Detects various account types, including username/password

Cons:

  • Limited visibility into personal accounts
  • Actual capabilities vary by vendor

Questions to ask of vendors:

  • What metadata does your method of email discovery collect about app usage? (Beware of solutions that are limited to email headers.)
  • Can you discover other details about AI use, like spend?

API connections with SaaS apps (for example, SSPM solutions)

SaaS Security Posture Management, or SSPM solutions typically work via direct API connections with specific SaaS apps and can provide detailed visibility into app-to-app integrations, and app configurations. For example, if you connect Salesforce to your SSPM provider, you could then potentially see all of the app-to-app connections (via OAuth or API) that have been granted between Salesforce and any other SaaS tool, including AI tools.

However, these tools do not offer true discovery given that you have to establish integrations between the SSPM tool and the apps you want to monitor with it. While this method can’t discovery “shadow” AI use, it is helpful for gaining visibility into integrations between your critical SaaS apps and AI tools so you can audit and revoke data sharing entitlements with AI tools.

How it works:

API connection between the SSPM solution and supported SaaS apps within your environment.

Pros:

  • Visibility into API and OAuth connections between AI tools and critical apps
  • Enables review of data sharing entitlements for AI tools
  • Continuous security monitoring for misconfigurations and risks

Cons:

  • Limited to known apps
  • API availability varies by SaaS provider
  • Higher effort to deploy relative to other options
  • Finite app coverage

Questions to ask of vendors:

How does your solution handle remediation of security findings when API automation is not available?

Summing it up: SaaS security solutions

As you’ve likely surmised, no single approach for AI discovery can give you the full picture of AI use, which is why most SaaS security solution providers offer a layered approach comprising more than one discovery method.

For example, Nudge Security combines email discovery, a browser extension, API connections, and SSO integration in order to address most AI discovery needs.

  • Email discovery provides broad historical coverage and can identify previously unknown AI tools without requiring prior knowledge of their existence.
  • Browser extensions offer real-time monitoring and intervention capabilities at the point of use.
  • API connections with critical SaaS apps deliver detailed insights into data sharing relationships between your core systems and AI tools.
  • SSO integration enriches your SaaS and AI inventory and supports SSO onboarding efforts

Together, these complementary methods provide both breadth (discovering the entire landscape of AI usage) and depth (understanding specific usage patterns and data flows), enabling security teams to implement appropriate controls while supporting legitimate business needs for AI adoption.

The emerging AI security market

Now that we’ve covered how SaaS security solutions are evolving to meet the AI discovery and governance challenge, we’ll circle back to the growing list of AI security point solutions.

For the most part, these tools leverage discovery methods similar to the browser-based discovery method described above. Some offer additional governance options like obfuscating sensitive info that users share in prompts (credit card numbers, SSNs, API keys, etc.)

The most important caveat to be aware of with these tools is that they are focused specifically on AI tools, not broader SaaS usage. And, as we’ve covered above, the lines are becoming very much blurred between what is an “AI tool” and what is a “SaaS tool” as SaaS vendors race to incorporate AI-enabled functionality. Additionally, integrations between SaaS tools and AI tools, as well as MCP servers introduce data security risks far beyond what a use might share in a prompt.

How it works:

AI security point solutions deploy advanced monitoring systems through browser extensions, endpoint agents, and API integrations with leading AI platforms to track and control how employees interact with AI tools.

Pros:

  • Deep visibility into prompt-level interactions
  • Real-time monitoring of specific AI tools
  • Specialized AI risk detection capabilities
  • Content filtering and sensitive data masking
  • AI-specific policy enforcement options
  • Protection against prompt injection attacks

Cons:

  • Limited support of AI tools
  • Fragmented visibility across the AI ecosystem
  • Often miss AI embedded within SaaS apps
  • Minimal historical data collection capabilities
  • Implementation complexity for distributed workforces
  • Separate from broader SaaS security governance
  • Focus on symptoms rather than comprehensive governance
  • Difficulty scaling with rapid AI tool proliferation

Questions to ask of vendors:

  • What specific AI security risks does your solution address? (prompt injection, data leakage, etc.)
  • How does your solution discover AI tools across the organization?
  • Can your solution detect AI embedded within existing SaaS applications?
  • How do you handle historical AI usage data versus forward-looking discovery?
  • What visibility do you provide into AI-to-SaaS integrations and data flows?
  • How does your solution balance security with employee productivity?
  • What governance capabilities do you offer beyond technical controls?
  • How does your solution integrate with our existing security stack?

Critical questions to ask about vendors’ AI discovery capabilities

We’ve covered a lot of points to consider. To help you with your evaluation process, here’s a condensed list of questions to ask of vendors regarding their AI discovery capabilities to assess how well they will meet your needs.

  • What AI discovery method(s) do you use?
  • Is your AI discovery method forward-looking only or can you discover AI assets created before your solution was deployed?
  • Does your AI discovery method rely on a known list of AI tools or can it dynamically identify new, never-before seen AI tools?
  • Which assets can your discovery method detect?
    • AI apps
    • User accounts
    • OAuth integrations between AI tools and other apps
    • API integrations between AI tools and other apps
    • AI in the supply chain of other SaaS providers
    • Data shared via AI prompts
    • Trends of AI tool adoption and usage patterns
    • Spend on AI tools
  • What insights can you provide on the AI tools that are discovered?
    • Security program details
    • Data training policy summaries
    • Breach histories
    • Data locality
    • Compliance attestations
  • Can you discover integrations between AI tools and your other business tools? If so, how?
  • Can you discover MCP server integrations?

AI discovery methods comparison chart

As you’ve probably surmised, there's no magic bullet when it comes to finding all the AI tools and integrations floating around your company. Each discovery approach has its strengths and blind spots. Understanding what works best will help you pick the right solution (or mix of solutions) to keep tabs on your AI landscape.

Here’s a table summarizing the discovery methods available, and the pros and cons:

Method Pros Cons
DIY AI hunting
  • Free (well, sort of)
  • Time-consuming
  • Quickly outdated
  • Incomplete answers spread across multiple sources
Network-based monitoring and endpoint agents (e.g. CASBs)
  • Monitors cloud service access
  • Applies security policies
  • Limited to corporate networks
  • Challenging for remote work
  • Resource intensive
  • Requires network proxies or endpoint agents
  • No historical discovery
Browser-based discovery*
  • Granular user activity data
  • Real-time interventions
  • Insights into usage frequency
  • Misses mobile and personal devices
  • Historical visibility limited to browser history
  • Requires browser plug-in
  • May have limited domain support
Email-based discovery
  • Broad discovery coverage
  • Finds unknown applications
  • Historical insight
  • Limited visibility into personal accounts
  • Actual capabilities vary by vendor
API connections
  • Visibility into integrations between AI tools and critical apps
  • Review of data sharing entitlements for AI tools
  • Continuous security monitoring for misconfigurations and risks
  • Limited to known apps
  • API availability varies by SaaS provider
  • Higher effort to deploy relative to other options
  • Finite app coverage

*This is the method typically employed by AI-specific point solutions

The ideal AI discovery solution

Effective AI discovery is crucial for modern organizations managing a distributed workforce and complex environments. Because each discovery method has its strengths and limitations, organizations should carefully evaluate their specific needs, infrastructure, and work environment when selecting an AI discovery solution.

The ideal solution should not only discover AI applications but also uncover where AI is embedded and integrated across the entire SaaS ecosystem. As the AI landscape continues to evolve and SaaS providers add AI-enabled capabilities, the line between what constitutes and “AI tool” vs. a “SaaS tool” will continue to dissolve.

By understanding the various discovery methods available and their respective trade-offs, teams can make informed decisions about which approach—or combination of approaches—best suits their needs now and can scale for the future. Remember that AI discovery is not a one-time exercise but an ongoing process that requires continuous monitoring and adaptation to keep pace with the evolving nature of AI use.

AI discovery with Nudge Security

Nudge Security takes a unique and comprehensive approach to AI discovery that overcomes the limitations of traditional methods. Our patented technology combines multiple discovery methods, with email-based discovery at its core, to provide unmatched visibility into AI (and SaaS) use. Here are the ways this method stands out from the rest:

  • Comprehensive coverage: While other solutions focus on narrow discovery methods, Nudge Security's multi-layered approach combines email analysis, API integrations, SSO connections, and a browser extension to provide the broadest possible visibility into AI use.
  • Automated discovery: Our solution works continuously in the background to identify new AI applications as soon as they appear in your environment, requiring minimal setup and maintenance and no prior knowledge of an apps existence.
  • Historical insight: Unlike future-looking solutions, Nudge Security uncovers AI accounts created in the past, giving you immediate visibility into your entire AI footprint, even apps introduced before you started using Nudge.
  • Rich context: Beyond just discovering AI apps, we provide detailed information about usage patterns, user roles, authentication methods, and security configurations to help you prioritize your efforts to govern AI use.
  • Scalability: Our approach doesn't require complex infrastructure, endpoint agents, or network monitoring tools, making it ideal for modern, distributed workforces.

Most importantly, Nudge Security's discovery capabilities serve as the foundation for a complete SaaS and AI security solution that helps organizations manage their entire SaaS attack surface, from discovery through governance with automated guardrails that guide your workforce to use SaaS and AI in safe, compliant ways.

“Having Nudge has significantly brought peace of mind because I don't have to go looking for a needle in a haystack anymore. This has been my dream that I've been looking for for a long time, for years.” —Leo C, IT team member at GLAAD

Related posts

Report

Debunking the "stupid user" myth in security

Exploring the influence of employees’ perception
and emotions on security behaviors