Our Head of Data, Michelle Urberg, co-authored this post with Chris Bendall (VP Strategy & Corporate Development at Research Solutions) on the future of Third Party Agentic Usage in the Society for Scholarly Publishing’s blog, the Scholarly Kitchen. We’ve reproduced their article below, or you can read the Scholarly Kitchen’s original version.

Did you know that widely used metrics only measure first-party usage (the publisher platform) and only the usage generated by humans? This excludes a rapidly growing body of content usage generated by AI (artificial intelligence) tools, both first‑party AI tools embedded in publisher platforms (e.g., Wiley) and especially by third-party agentic platforms (e.g., Claude, ChatGPT).

That means, there’s an elephant-sized gap in most usage reports.

Third-Party Agentic Usage is the Direction of Travel

We are all feeling the ground shift underneath our feet in libraries and publishing. We don’t yet know the full extent of how agentic tools will change our industry. But we do know OpenAI’s “low-key research project” (aka, ChatGPT) has fundamentally uprooted research and teaching practices. Since 2023, this has been widely discussed, including in an Inside Higher Ed Op Ed and many subsequent opinion pieces.

Third-party agentic tools like ChatGPT, Claude, Gemini, and Perplexity, as well as scholarly content-specific tools like Consensus, Elicit, Scite, and others, can quickly fetch, sort, and organize information that would otherwise take humans weeks or months to collate. In some cases, this means users bypass traditional abstract databases, library catalogs, and full-text content platforms. AI outputs, including summaries, snippets, and citations, are deemed “good enough” — so much so that users are increasingly satisfied with the AI output and do not seek out a publisher platform for the full-text content.

The result is a dramatic drop in referrer traffic to full-text platforms and a dramatic rise in the “zero-click search,” leading to the “zero-click result.” A zero-click search happens when a user gets what they need from an AI-generated output and does not need to visit a content platform for more information. A zero-click result is the content of that search, which could include a full-text snippet, an abstract, or other bibliographic information about a resource.

Unsurprisingly, then, academic publishers are alternately panicking or fighting against the “zero-click search” and “zero-click result,” as well as their impacts on usage reporting (zero usage logged). Zero usage leaves both publishers and librarians in the dark about what content is valuable to their customers and patrons. In teaching and research contexts, this means that it is that much harder to make data-driven purchasing decisions and therefore harder to decide what books, journals, and other content to buy. This is a publisher’s worst nightmare and a librarian’s headache.

But there is good news! Usage has not vanished! Rather, it has been displaced. It has been rerouted to agentic platforms and transformed into different kinds of searches and content engagement. Wiley, Scite (full disclosure, one author of this post is associated with Scite), and Cashmere (as an infrastructure and rights-management provider rather than a discovery platform) have developed infrastructure to address such “zero-click searches” and the associated loss of usage metrics. In each case, they are leveraging API endpoints and especially the Model Context Protocol (MCP), developed by Anthropic in 2024, which rapidly emerged as an open standard for enabling agentic AI tools to interact with third-party tools, content repositories, and datasets in a structured, replicable, and auditable manner.

The “chunked” and indexed content is retrieved and assessed, via MCP, by the AI tool to determine its relevance for a query. How content is chunked can differ; for instance, Scite focuses on citations and their context, while Cashmere is oriented around concepts and their context. In addition, the business models for the services are rapidly evolving in relation to both publishers and customers, and we can expect different approaches to appear and be tested in the next few months.

These “chunks” of content are the future of usage tracking. Steve Smith argues that as AI shifts discovery from documents to answers, usage measurement (including COUNTER) will need to evolve beyond PDF downloads toward tracking machine-consumable “knowledge objects” with explicit context and provenance as the new unit of scholarly use. This is the moment to be proactive about managing this change. Our task now is to find where that usage has gone and reclaim it, so that libraries and publishers can continue to demonstrate value to patrons and customers. Collaboration with service providers developing agentic tools is the key, as they are now intermediaries between users and the libraries and publishers who provide them with content.

The Basics: How Do AI Tools Interact with Scholarly Content?

When a user queries an agentic AI tool (described also here Discovery #1: How AI agents transform researcher inputs), which is leveraging a source of indexed scholarly content or citations (either via API or MCP), it typically expands the query into several additional search strategies to ensure it finds the most relevant sources. The LLM then refines the number of sources by relevance to the human query and its expanded search strategies and provides a summary for the user.

Two key usage points come out of the generative AI search strategies:

  • AI Reads: Automated AI search related usage driven by the AI as it acts upon the expanded search strategies it has defined as a result of a single human query.
  • AI Citations: The refined list of sources upon which the summaries are based, ideally with links back to the Version of Record or other source document.

flow chart showing steps from a human entering a query on the aurora borealis through the AI processing it and including references in its answer

While generative and agentic AI searches are important for understanding the use of a given tool, the publications consulted (through abstracts and bibliographic information), and summaries of the cited sources are the content engagement points that libraries and publishers need most for tracking usage. As more publishers are indexed by platforms and initiatives, like Scite or Wiley’s Scholar Gateway, the more items can be counted and useful metrics generated.

The Challenge with Existing Metrics

In scholarly publishing, existing metrics for recording content usage are derived using the COUNTER Code of Practice, which presumes that the user is engaging directly with a first-party content platform (including library discovery services and content aggregators), not a third-party content provider or an agentic tool. COUNTER excels at defining content usage in first-party content providers by standardizing the types of user interactions, controlling vocabulary for naming content types, access types, and access methods. Because everyone uses the same metrics, usage can be compared across platforms. Without standard metrics and vocabularies, comparable usage is impossible to generate.

COUNTER has developed long-tested metrics, which are assigned to one of four types of interactions: searches, views of content info (investigations), views of full content (requests), and turnaways (denials). Imagine your own engagement with finding electronic books, journals, or media. You likely begin with a search, follow with clicks on desired items, and then conclude by playing a video or reading the full text (assuming you have access and aren’t turned away). This pattern — search, discovery, view — is repeated over and over again as users engage directly with a platform and its content.

So, What About COUNTER and Third-party Agentic Usage?

Usage generated in third-party platforms like ChatGPT, Claude, Scite, and others is currently out of scope for COUNTER reporting because the standard presumes a user is engaging directly with a first-party content provider. Since AI tools are starting to provide the same basic experience for human users as traditional content and discovery platforms — namely, search, investigate, view/denial — why not extend COUNTER metrics to measure third-party agentic-generated usage?

As a first step, COUNTER is exploring the addition of a new access method to the Code of Practice to accommodate first-party agentic usage (usage generated by AI tools inside of a publisher platform). We think usage tracking must be extended even further to count usage generated by third-party agentic tools.

Let’s Help Usage Metrics Catch Up to Current Technology

We know that researchers are using AI tools for more and more pieces of the research life cycle and that they want tools to provide credible results. Reliable third-party agentic platforms are not a passing trend; they are the direction of travel. One of the ways to increase credibility for researchers and libraries is to develop standard, auditable metrics to track usage generated in these platforms.

In order to make steps toward a future with standardized third-party usage metrics, we recommend that publishers and libraries take the following steps toward a future with standardized third-party usage metrics:

Track what we can: investigations, requests, and denials

The standard metadata collected for abstract and summary views (investigations), full-content views (the request), and denials (as defined in the Code of Practice) can immediately be used to measure agentic-generated usage. If the agentic tool is providing a citation or a full-text view of the content, these can be identified as an investigation or a request, respectively. If the agentic platform has full-text availability, but the user does not have licensed access, the denial should also be available for reporting.

Define what constitutes a search generated by a human vs an agent

Tracking searches generated by agentic tools is hard. As AI-assisted search becomes more prevalent across library discovery services and in publisher-owned platforms, fewer and fewer searches will be characterized as what COUNTER calls “searches regular” or “searches platform” (i.e., searches generated by a human user). AI-assisted searches set up multiple queries for each question posed to an agentic tool. Should we only count those human queries and not the AI-assisted ones? Or should we count both?

Accommodate knowledge objects by augmenting existing COUNTER metrics

The extraction of knowledge objects in the form of text chunks (snippets and/or individual pages) and agentic writing of summaries needs to be defined. COUNTER metrics can be extended to accommodate these usage points. Knowledge objects are a kind of full-text, but not the whole text. And a user could be presented with a portion of an article without seeing the full text or having access to it. These textual “sound bites” should be given some recognition, not unlike a quotation with citation in a paper. Publishers, at the very least, will want to get this usage counted to track how an agentic tool bolsters their content usage.

Count AI content usage that is leveraged to generate AI output

AI-generated outputs, such as summaries, are ephemeral, typically lacking a standard and persistent identifier. However, the indexed references analysed by an AI tool in answer to an initial human query (“AI Reads”) and the subset subsequently included in a summary (“AI Citations”) can and should be measured. This represents the rapidly growing source of usage (and a measure of the value and impact of the content) currently missed by “zero-click searches.”

Conclusion

We do not have all of the solutions today, but we can see the direction of travel: agentic tools are not going away for libraries and publishing. The sooner we come together to figure out how to harness the elephant of third-party agentic usage, the better we will be able to demonstrate the value of content for both libraries and publishers.

 

Post-Script on COUNTER Metrics

As indicated above, third-party usage sits at the edge of COUNTER metrics at present. However, we would like to highlight a couple of exceptions. These mechanisms for tracking usage are less commonly used, but are part of an auditable standard.

  • Usage generated by text and data mining (TDM) is one type of usage not generated by humans. TDM is a specific harvesting mechanism agreed upon by an organization and a library to download large amounts of content for research purposes. It requires setting up specific rules around harvesting data (see the COUNTER CoP Section 7.10).
  • Distributed usage logging (DUL) is the other provision for tracking third-party usage, but it still relies on humans generating the usage and not agentic systems. This model has had little uptake in the industry, suggesting that we need to find another way to bring third-party usage into analytics reporting. For more on DUL see COUNTER CoP Section 6.3).