Chethan Parameshwara

I am a researcher at Meta, where I develop real-time diffusion models for personalized AI assistants.

Prior to joining Meta, I was a Member of Technical Staff at WaveForms AI, focusing on multimodal diffusion models. Earlier, as a researcher at Amazon, I was a core contributor to Nova foundational models for multimodal generation.

I earned my Ph.D. in Neuroscience and Artificial Intelligence from the Neuroscience and Cognitive Science (NACS) program at the University of Maryland, College Park (UMD). My PhD research focused on video understanding and 3D computer vision for autonomous vehicles.

Email  /  GitHub  /  Google Scholar  /  LinkedIn  /  X

Publications·Experiences·Teaching·Volunteering
Publications

My selected publications are listed here. The complete list of publications can be seen from my Google Scholar page.

The Amazon Nova family of models: Technical report and model card
Amazon Artificial General Intelligence, 2024
technical report / video

Core contributor to Amazon Nova Reel, an advanced video generation model within the Amazon Nova foundation suite, designed to produce high-quality, customizable outputs with fine-grained motion control.

Towards Visual Foundational Models of Physical Scenes
Chethan Parameshwara*, Alessandro Achille*, Matthew Trager, Xiaolong Li, Jiawei Mo, Ashwin Swaminathan, CJ Taylor, Dheera Venkatraman, Xiaohan Fei*, Stefano Soatto* (* equal contribution)
arXiv, 2023

We describe a first step towards learning general-purpose visual representations of physical scenes using only image prediction as a training criterion.

Towards an Improved Hyperdimensional Classifier for Event-Based Data
Neal Anwar, Chethan Parameshwara, Cornelia Fermüller, Yiannis Aloimonos
CISS, 2023
project page / arXiv

We present a novel bipolar HD encoding mechanism designed for encoding spatio-temporal data of an asynchronous event camera.

DiffPoseNet: Direct Differentiable Camera Pose Estimation
Chethan Parameshwara, Gokul Hari, Cornelia Fermüller, Nitin J. Sanket, Yiannis Aloimonos
CVPR, 2022
project page / arXiv / video

We propose a novel differentiable programming technique for camera pose estimation in autonomous driving and flying scenarios.

SpikeMS: Deep Spiking Neural Network for Motion Segmentation
Chethan Parameshwara*, Simin Li*, Cornelia Fermüller, Nitin J. Sanket, Matthew S. Evanusa, Yiannis Aloimonos (* equal contribution)
IROS, 2021
project page / code / arXiv / video

We propose a bio-inspired neural network for the motion detection problem using an asynchronous event camera.

NudgeSeg: Zero-Shot Object Segmentation by Repeated Physical Interaction
Chahat Deep Singh*, Nitin J. Sanket*, Chethan Parameshwara, Cornelia Fermüller, Yiannis Aloimonos (* equal contribution)
IROS, 2021
project page / arXiv / video

We present a novel approach to segment never-seen objects by physically interacting with them and observing them from different views.

0-MMS: Zero-Shot Multi-Motion Segmentation With A Monocular Event Camera
Chethan Parameshwara, Nitin J. Sanket, Chahat Deep Singh, Cornelia Fermüller, Yiannis Aloimonos
ICRA, 2021
project page / code / arXiv / video

We propose a hybrid solution to multi-object motion segmentation using a combination of model-based and deep learning approaches with minimal prior knowledge.

EVPropNet: Detecting Drones By Finding Propellers For Mid-Air Landing And Following
Nitin J. Sanket, Chahat Deep Singh, Chethan Parameshwara, Cornelia Fermüller, Guido de Croon, Yiannis Aloimonos
RSS, 2021
project page / code / arXiv / video

We present a deep learning-based solution for detecting propellers (to detect drones) for mid-air landing and following.

EVDodgeNet: Deep Dynamic Obstacle Dodging with Event Cameras
Chethan Parameshwara*, Nitin J. Sanket*, Chahat Deep Singh, Ashwin V. Kuruttukulam, Cornelia Fermüller, Davide Scaramuzza, Yiannis Aloimonos (* equal contribution)
ICRA, 2020
project page / code / arXiv / video

We present the first deep learning based solution for dodging multiple dynamic obstacles on a quadrotor with a single event camera and onboard computation.

Event-based Moving Object Detection and Tracking
Anton Mitrokhin, Cornelia Fermüller, Chethan Parameshwara, Yiannis Aloimonos
IROS, 2018
project page / code / arXiv / video

We present a novel motion compensation approach for moving object tracking with an asynchronous event camera.

Experiences
PontTuset

Meta
Aug, 2025 -- Present
Researcher

Research on real-time diffusion models for personalized AI assistants.

PontTuset

WaveForms AI
May, 2025 -- Aug, 2025
Member of Technical Staff

Research on real-time multimodal diffusion models for audio-video generation.

PontTuset

Amazon
Aug, 2022 -- May, 2025
Researcher

Research on video diffusion models (text-to-video, image-to-video, video-to-video) and 3D computer vision (NeRF, Gaussian Splatting, Neural Rendering, and SLAM).

PontTuset

University of Maryland, College Park (UMD)
Aug, 2017 -- Aug, 2022
Graduate Research Assistant with Prof. Yiannis Aloimonos and Dr. Cornelia Fermüller

Research on video understanding (object detectiona and segmentation, motion segmentation) and 3D computer vision (pose estimation, SLAM) for autonomous vehicles.

PontTuset

SRI International (formerly Stanford Research Institute)
Summer, 2021
Research Intern with Dr. David Zhang and Michael Piacentino

Research on few-shot deep learning and gradient-free learning algorithms.

Teaching
panoroma Teaching Assistant, CMSC733 : Geometric Computer Vision

Teaching Assistant, CMSC426 : Computer Vision

Teaching Assistant, CMSC434 : Human Computer Interaction
Volunteering
gsglogo Reviewer: NeurIPS, ICLR, CVPR, ECCV, RAL, ICRA, IROS, IEEE Sensor Journal

NACS Representative, UMD Graduate Student Government (2020-2021)

Co-Chair, NACS Grant Review Committee (2019-2022)

Staff Member, Neuromorphic Cognition Engineering Workshop (July 2018)

Thanks to Jon Barron for this minimalist website template.