Research Group

Our group sits at the nexus of computer vision, graphics and AI, working with the creative industries to design practical algorithms and open tools for trustworthy media and content creation. We focus on content authenticity and media provenance technologies, as well as representation learning for visual search and content discovery. Our work within the DECaDE centre explores how AI and distributed ledgers (DLT) can enable fair, resilient creative supply chains that support consent, attribution and monetisation in the age of generative AI. We collaborate closely with industry and policymakers, and aim to translate research into standards and real-world deployments. If you’re interested in a PhD aligned to these themes, informal enquiries and visits are welcome.

DECaDE research team at 2024 plenary meeting Presenting DECaDE research at the House of Lords

DECaDE: UKRI/EPSRC Next Stage Centre for the Decentralised Digital Economy

I am founder and director of DECaDE, a multi-disciplinary research centre exploring the future digital economy, with a focus on the creative industries. We live in a decentralised digital economy where anyone can produce or consume digital goods and services (e.g. Uber, AirBnB), yet this activity is mediated by centralised platforms with closed governance.

DECaDE investigates how data-centric technologies such as AI and Distributed Ledger Technology (DLT) can enable decentralised platforms that support fairer governance, build trust and resilience, and open new models of value creation. A core theme is digital supply chains, using decentralised technologies to track the provenance of digital assets (or digital twins of physical assets).

Our current emphasis is on the creative supply chain: provenance of media to establish authenticity (combatting misinformation and deepfakes) and to improve attribution and reward for creators - issues made more urgent by Generative AI.

DECaDE not only explores technological aspects but also incorporates researchers from the Surrey Business School and Surrey Centre for Cyber Security, alongside academics from Edinburgh’s School of Law and Institute for Design Informatics, reflecting the deeply socio-technical nature of this space. DECaDE partners with the Digital Catapult, whose DLT Field Labs help scale our prototypes into real-world trials. Read more on the DECaDE website and in the DECaDE Impact brochure (2024).

Content Authenticity and Media Provenance

Our group addresses the societal threat of fake news and misinformation. While many researchers focus on deepfake detection, not all AI-generated content is deceptive, and building detectors is an arms race with GenAI. Human rights organizations note that most visual misinformation is not AI-generated but genuine content miscontextualized or misattributed. For this reason, we see media provenance as crucial—enabling users to trace content origins and make informed trust decisions. My research develops technologies for provenance (secure metadata, watermarking, fingerprinting, distributed ledgers) and contributes to international standards such as C2PA for communicating provenance.

ARCHANGEL (2017-19) was one of my earliest media provenance projects, working with National Archives around the world to help tamper-proof content by storing visual provenance information on blockchain. This work was highlighted by UKRI/EPSRC as a highlight of its 10 year Digital Economy research portfolio. Other projects included TAPESTRY (identity provenance, 2016-2018) and CoMEHeRe (an early exploration of healthcare data monetization based on provenance, 2017-2019).

Since joining Adobe Research in 2019, I co-founded the Content Authenticity Initiative (CAI), an industry coalition now exceeding 5000 members (2025). CAI's technical framework evolved into a cross-industry standard (C2PA; Coalition for Content Provenance and Authenticity). I have been involved in the C2PA technical working group from the outset, chairing two task forces focused upon watermarking and blockchain. These technologies enable C2PA metadata embedded within assets (referred to as Content Credentials) to become more durable i.e. survive content redistribution, particularly through social media platforms that strip metadata. Our joint Adobe–Surrey work has produced perceptual fingerprinting [e.g. OscarNet, ICCV 2021; ICN CVPRW 2021], open-source watermarking [e.g. RoSteALS, CVPRW 2023; TrustMark, ICCV 2025], methods to summarize provenance [e.g. VIXEN, ICCV 2023; ImProvShow, BMVC 2025], and studies its value to users [ACM C&C 2025].

I have been involved in policy and regulatory discussions around misinformation and online harms. As part of the Royal Society Pairing Scheme, I embedded within the UK Department of Science Innovaton and Technology (DSIT) disinformation team and have also presented to DARPA, the US National Academy of Science and the European Commission among others to help inform policymakers about the emerging technical landscape around media provenance. Our reseach papers are often cited by public bodies reporting in the media authenticity space e.g. the NSA/NCSC and World Privacy Forum reports on C2PA.

Sample of relevant publications

“TrustMark: Robust Watermarking and Watermark Removal for Arbitrary Resolution Images”
T. Bui, S. Agarwal, J. Collomosse
IEEE International Conference on Computer Vision (ICCV), 2025 pdf bib
“To Authenticity, and Beyond! Building Safe and Fair Generative AI upon the Three Pillars of Provenance”
J. Collomosse, A. Parsons
IEEE Computer Graphics and Applications (IEEE CG&A), 2024 pdf bib
“ImProvShow: Multimodal Fusion for Image Provenance Summarization”
A. Black, J. Shi, Y. Fan, J. Collomosse
British Machine Vision Conference (BMVC), 2025 pdf bib
“RoSteALS: Robust Steganography using Autoencoder Latent Space”
T. Bui, S. Agarwal, N. Yu, J. Collomosse
CVPR Workshop on Media Forensics (CVPRW), 2023 pdf bib
“Content Authenticities: A Discussion on the Values of Provenance Data for Creatives and Their Audiences”
C. Moruzzi, E. Tallyn, F. Liddell, B. Dixon, J. Collomosse, C. Elsden
ACM Creativity and Cognition (C&C), 2025 pdf bib
“OSCAR-Net: Object-centric Scene Graph Attention for Image Attribution”
E. Nguyen, T. Bui, V. Swaminathan, J. Collomosse
IEEE International Conference on Computer Vision (ICCV), 2021 pdf bib
Interoperable Watermarking for C2PA
Durable Content Credentials (C2PA)
ARCHANGEL Blockchain Participants
Deep Image Comparator (CVPR WMF 2021)
TrustMark Performance Comparison (ICCV 2025)

Artistic Stylization and Style Representation

My PhD research explored AI and Computer Vision to create tools to transfer artistic styles onto photographs and video (then referred to as Non-Photorealistic Rendering). After the advent of deep learning and now Generative AI, a core research activity of has been to develop accurate, fine-grained representations of style both for content creation (e.g. neural style transfer) and for visual search (e.g. style descriptors). Our ALADIN style representation [ECCV, 2022] has been widely adopted in research for both tasks and were commercialized to drive style search in the Behance platform. More recently we have explored ways to condition generative AI (diffusion) models using ALADIN [NeAT, ECCV 2024; DIFF-NST, ECCVW 2024] and also to train vision language models to describe style [StyleBabel, ECCV 2022].

Sample of relevant publications

“ALADIN: All Layer Adaptive Instance Normalization for Fine-grained Style Similarity”
D. Ruta, S. Motiian, B. Faieta, Z. Lin, H. Jin, A. Filipkowski, A. Gilbert, J. Collomosse
IEEE International Conference on Computer Vision (ICCV), 2021 pdf bib
“StyleBabel: Artistic Style Tagging and Captioning”
D. Ruta, A. Gilbert, P. Aggarwal, N. Marri, A. Kale, J. Briggs, C. Speed, H. Jin, B. Faeita, A. Filipkowski, Z. Lin, J. Collomosse
European Conference on Computer Vision (ECCV), 2022 pdf bib
“DIFF-NST: Diffusion Interleaving For deFormable Neural Style Transfer”
D. Ruta, G. Canet Tarres, A. Gilbert, E. Shechtman, N. Kolkin, J. Collomosse
European Conference on Computer Vision (ECCV), 2024 pdf bib
“PARASOL: Parametric Style Control for Diffusion Image Synthesis”
G. Canet Tarres, D. Ruta, T. Bui, J. Collomosse
CVPR Workshop on Women in Computer Vision (WiCV), 2024 pdf bib
“SceneComposer: Any-Level Semantic Image Synthesis”
Y. Zeng, Z. Lin, J. Zhang, Q. Liu, J. Collomosse, J. Kuen, V. M. Patel
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023 pdf bib
ALADIN style representation
NeAT Neural Style Transfer
PARASOL fine-grained style control over diffusion GenAI
LiveSketch Network Architecture (CVPR 2019)
Sketch-based visual search using GF-HOG (CVIU 2013)

Sketch-based Visual Search

Our group was among the earliest to explore sketch based visual search of image and video databases. We developed the first technique for sketch based video search [ICCV 2009] and popular datasets and descriptors for sketch based image search (noably GF-HoG [CVIU, 2013]). We explored other interesting forms of visual search in this domain, such as sketch based choreography search as part of the AHRC funded Digital Dance Archives (DDA) project. Several PhDs were run in this space (e.g. Rui Hu, Stuart James, Tu Bui) over those years exploring novel descriptors and indexing strategies for scalable sketch based search. Our group is no longer active in sketch based research however the Sketch-X group run by Prof. Song at Surrey remains very active in this space.

Sample of indicative publications

“LiveSketch: Query Perturbation for Guided Sketch-based Visual Search”
J. Collomosse, T. Bui, H. Jin
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019 pdf bib
“A Performance Evaluation of the Gradient Field HOG Descriptor for Sketch Based Image Retrieval”
R. Hu, J. Collomosse
Computer Vision and Image Understanding (CVIU), 2013 pdf bib
“Visual Sentences for Pose Retrieval over Low-resolution Cross-media Dance Collections”
R. Ren, J. Collomosse
IEEE Transactions on Multimedia, 2012 pdf bib
“Storyboard sketches for content based video retrieval”
J. Collomosse, G. McNeill, Y. Qian
IEEE International Conference on Computer Vision (ICCV), 2009 pdf bib

4D Performance Capture

Surrey has a long-standing track record of performance capture research in collaboration with the UK creative sector, such as the BBC, DNEG, etc. During my earlier years at Surrey I worked extensively on this topic, helping to develop novel technologies for markerless 4D surface capture and multi-view video matteing and segmentation. We were one of the first groups to fuse volumetric and IMU based data for markerless motion capture, releasing a popular dataset [Total Capture, ECCV 2018] in this domain. These projects were primarily funded by InnovateUK (e.g. REFRAME) and the EU FP7 and Horizon 2020 programmes (e.g. RE@CT). During this time we were also developing the University’s links with Guildford’s local video gaming industry, and I co-founded a regional conference series ‘G3’ (2015-2018) to help promote this collaboration.

Sample of relevant publications

“Hybrid Skeletal-Surface Motion Graphs for Character Animation from 4D Performance Capture”
P. Huang, M. Tejera, J. Collomosse, A. Hilton
ACM Transactions on Graphics (TOG), 2015 pdf bib
“Volumetric performance capture from minimal camera viewpoints”
A. Gilbert, M. Volino, A. Hilton, J. Collomosse
European Conference on Computer Vision (ECCV), 2018 pdf bib
“Real-Time Multi-person Motion Capture from Multi-view Video”
C. Malleson, J. Collomosse, A. Hilton
International Journal of Computer Vision (IJCV), 2020 pdf bib
4D performance capture
4D performance capture