
Using a Multi‑Modal Content Graph and AI across your sports media delivers the ultimate sports knowledge hub – unlocking massive latent value in video, audio, imagery, statistics, and editorial content.
Modern sports organizations sit on vast archives of matches, highlights, interviews, commentary, statistics, and written analysis. Yet, most of this content remains siloed, under‑described, and difficult to explore beyond simple keyword search. This is where Data Graphs – an AI‑powered, enterprise‑grade, multi‑modal knowledge graph – fundamentally changes what is possible.
This article expands on the demo outlined below and shows how Data Graphs can be used to explore, analyze, and increase the value of sports media by combining video, sports data, and AI into a single, connected knowledge fabric.
The primary mechanism for unlocking value is to join and align sports media with sports data metadata using a structured domain model (ontology) for both media and sport.
For this work, we use the IPTC Sports Schema, a well‑designed, open, and extensible model for representing:
In Data Graphs, the IPTC Sports Schema is combined with:
With respect to sports data enrichment, there are two complementary strategies:
The result is a single, queryable, explainable graph that connects what happened, who was involved, and where it appears in the media.
Using 2015–2017 Premier League and Champions League video and data from SoccerNet, we created a practical demonstration of how Data Graphs can deliver the ultimate AI‑powered sports knowledge hub.
This demo combines:
Data Graphs is much more than a graph database. Its native handling of video, audio, text, and embeddings makes it a true content graph for:
Users can ask natural‑language questions such as:
Behind the scenes, these questions are answered by combining semantic vector search with precise graph queries over structured sports data.
At the core of the system is a clean, explicit domain model. The IPTC Sports Schema provides the backbone for sporting concepts, while CreativeWorks and Video ontologies describe the media itself.
This allows Data Graphs to natively represent relationships such as:
Because the model is explicit, it remains explainable, extensible, and portable across competitions and sports.
For the demo, we used:
SportsMonks aligns closely with the IPTC model, making ingestion straightforward – but any sports data provider can be used. Data Graphs is intentionally schema‑flexible while still enforcing semantic consistency.
A critical step is linking structured events (goals, fouls, substitutions) to time‑coded moments in video.
Each in‑game action becomes a first‑class node in the graph, connected to:
This instantly turns raw match footage into a searchable, explorable library of meaningful moments.
Because everything lives in a graph, powerful queries become simple and expressive:
Data Graphs supports GQL/OpenCypher, making it easy for developers, analysts, and AI agents to programmatically explore the data.
Beyond code, Data Graphs provides visual graph exploration, allowing users to:
This is invaluable for editors, analysts, and archivists who want insight without writing queries.
Data Graphs includes a Model Context Protocol (MCP)‑based AI architecture that enables multiple specialized agents to operate on the same graph.
Agents can be configured to:
Unlike vector‑only RAG systems, Data Graphs delivers true hybrid GraphRAG out of the box:
This means an AI assistant can seamlessly combine statistical accuracy with rich contextual understanding. The Data Graphs native AI assistant operates using this architecture.
One of the most powerful features is the ability to render views contextually. Instead of returning raw JSON or text, Data Graphs can render rich, contextual views for:
These views dynamically pull together data, media, and AI insights from across the graph – creating human‑friendly representations that can be embedded into applications, dashboards, or editorial tools.
In the example rendered view shown above, a single match concept (English Premier League 2015–2016, Chelsea v Aston Villa) becomes the organising context for:
Each timeline entry is not just a label, but a graph entity linked to precise video moments, participants, and statistics. Users can move seamlessly from a goal in the timeline to the video clip, to the player profile, and out into the wider competition graph.
This approach replaces hard‑coded pages with schema‑driven, graph‑powered views that adapt automatically as new data, competitions, or media are added.
AI video analysis services such as Bitmovin's Content Analysis AI can be integrated to automatically:
This metadata flows directly into the graph, continuously enriching the knowledge base and reducing manual annotation costs.
A multi‑modal sports knowledge graph enables:
In short, it turns passive sports media archives into active, intelligent assets.
The IPTC Sports Schema is sport‑agnostic and competition‑agnostic. The same approach applies to:
Simply feed Data Graphs with your media and sports data, and the platform does the rest – connecting, enriching, and activating your content with AI.
Data Graphs shows how multi‑modal knowledge graphs are no longer experimental technology – they are a practical foundation for the next generation of sports media, analytics, and fan experiences.
See how it works - get in touch for a demo!