<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Sotirios Chatzis | MSc in Electronics and Technology</title><link>https://mscest.cut.ac.cy/author/sotirios-chatzis/</link><atom:link href="https://mscest.cut.ac.cy/author/sotirios-chatzis/index.xml" rel="self" type="application/rss+xml"/><description>Sotirios Chatzis</description><generator>Wowchemy (https://wowchemy.com)</generator><language>en-us</language><lastBuildDate>Thu, 01 Jan 2026 00:00:00 +0000</lastBuildDate><image><url>https://mscest.cut.ac.cy/media/logo_hude1662fe81542519856cdd9b507606f3_856625_300x300_fit_lanczos_3.png</url><title>Sotirios Chatzis</title><link>https://mscest.cut.ac.cy/author/sotirios-chatzis/</link></image><item><title>CMAD: Conditional Modeling-Adapter Diffusion for Video Super-Resolution</title><link>https://mscest.cut.ac.cy/publication/2025_chengzhang_wang_cmad-video-super-resolution/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://mscest.cut.ac.cy/publication/2025_chengzhang_wang_cmad-video-super-resolution/</guid><description>&lt;p>Video super-resolution (VSR) aims to restore high-resolution videos from low-resolution inputs, enhancing visual quality under real-world degradation. Existing diffusion-based VSR methods often rely on specially designed network architectures or text prompts as conditional inputs, which limits their flexibility and applicability&amp;mdash;especially in scenarios where explicit text descriptions are unavailable. To tackle this challenge, we propose a Conditional Modeling ADapter that enables zero-modification reuse of pre-trained text-to-video diffusion models for the VSR task, called CMAD. The adapter transforms low-resolution video features into pseudo-text tokens via a pre-trained Vision Transformer (ViT) and a lightweight adapter module, allowing these features to be injected as encoder hidden states, which serve as the main conditioning interface within the diffusion model. Unlike previous methods that require redesigning or retraining task-specific backbone architectures, our design enables the model to interpret video inputs directly through its original language conditioning interface&amp;mdash;without any modification to the pre-trained diffusion network. Experiments on several VSR benchmarks, including REDS, UDM10, and VID4, demonstrate that our method achieves reconstruction performance comparable to specialized super-resolution models, without text guidance. Moreover, CMAD provides a simple and efficient solution for adapting large-scale pre-trained diffusion models to video super-resolution tasks. The source code and models will be released.&lt;/p></description></item></channel></rss>