Abstract
Artificial Intelligence (AI) research has grown at a fast pace in recent years, introducing a variety of new methods (e.g., Large Language Models), specialized datasets (e.g., ImageNet), and emergent tasks (e.g., multi-modal text generation). Traditional bibliometric analyses shed light on publication volumes and co-authorship patterns but often fail to capture the specific concepts that underlie these collaborations. To address this gap, we propose a two-tiered approach that merges attention-based concept mining with co-authorship network analysis. First, we integrate frequency-based and dictionary-based techniques, guided by anchor-phrases, to train a neural network that identifies not only high-frequency AI terms but also rare, high-impact concepts. By leveraging attention maps from a pre-trained transformer, we reduce noise—distinguishing genuine new ideas like “Vision Transformer” from random token clusters. Second, we situate these concepts in temporal co-authorship networks, applying community detection to reveal where innovations emerge, how they travel across research subfields, and what collaboration patterns accelerate or inhibit their diffusion. This synthesis provides a robust framework for research methods, enabling richer inquiries into how AI knowledge diffuses and reshapes scientific collaboration.