|
Chethan Parameshwara
I am a researcher at Meta, where I develop real-time diffusion models for personalized AI assistants.
Prior to joining Meta, I was a Member of Technical Staff at WaveForms AI, focusing on multimodal diffusion models. Earlier, as a researcher at Amazon, I was a core contributor to Nova foundational models for multimodal generation.
I earned my Ph.D. in Neuroscience and Artificial Intelligence from the Neuroscience and Cognitive Science (NACS) program at the University of Maryland, College Park (UMD). My PhD research focused on video understanding and 3D computer vision for autonomous vehicles.
Email  / 
GitHub  / 
Google Scholar  / 
LinkedIn  / 
X
|
|
|
Publications
My selected publications are listed here. The complete list of publications can be seen from my Google Scholar page.
|
|
|
The Amazon Nova family of models: Technical report and model card
Amazon Artificial General Intelligence, 2024
technical report /
video
Core contributor to Amazon Nova Reel, an advanced video generation model within the Amazon Nova foundation suite, designed to produce high-quality, customizable outputs with fine-grained motion control.
|
|
|
Towards Visual Foundational Models of Physical Scenes
Chethan Parameshwara*,
Alessandro Achille*,
Matthew Trager,
Xiaolong Li,
Jiawei Mo,
Ashwin Swaminathan,
CJ Taylor,
Dheera Venkatraman,
Xiaohan Fei*,
Stefano Soatto* (* equal contribution)
arXiv, 2023
We describe a first step towards learning general-purpose visual representations of physical scenes
using only image prediction as a training criterion.
|
|
|
Towards an Improved Hyperdimensional Classifier for Event-Based Data
Neal Anwar,
Chethan Parameshwara,
Cornelia Fermüller,
Yiannis Aloimonos
CISS, 2023
project page /
arXiv
We present a novel bipolar HD encoding mechanism designed for encoding spatio-temporal data of an asynchronous event camera.
|
|
|
DiffPoseNet: Direct Differentiable Camera Pose Estimation
Chethan Parameshwara,
Gokul Hari,
Cornelia Fermüller,
Nitin J. Sanket,
Yiannis Aloimonos
CVPR, 2022
project page /
arXiv /
video
We propose a novel differentiable programming technique for camera pose estimation in autonomous driving and flying scenarios.
|
|
|
SpikeMS: Deep Spiking Neural Network for Motion Segmentation
Chethan Parameshwara*,
Simin Li*,
Cornelia Fermüller,
Nitin J. Sanket,
Matthew S. Evanusa,
Yiannis Aloimonos (* equal contribution)
IROS, 2021
project page /
code /
arXiv /
video
We propose a bio-inspired neural network for the motion detection problem using an asynchronous event camera.
|
|
|
NudgeSeg: Zero-Shot Object Segmentation by Repeated Physical Interaction
Chahat Deep Singh*,
Nitin J. Sanket*,
Chethan Parameshwara,
Cornelia Fermüller,
Yiannis Aloimonos (* equal contribution)
IROS, 2021
project page /
arXiv /
video
We present a novel approach to segment never-seen objects by physically interacting with them and observing them from different views.
|
|
|
0-MMS: Zero-Shot Multi-Motion Segmentation With A Monocular Event Camera
Chethan Parameshwara,
Nitin J. Sanket,
Chahat Deep Singh,
Cornelia Fermüller,
Yiannis Aloimonos
ICRA, 2021
project page /
code /
arXiv /
video
We propose a hybrid solution to multi-object motion segmentation using a combination of model-based and deep learning approaches with minimal prior knowledge.
|
|
|
EVPropNet: Detecting Drones By Finding Propellers For Mid-Air Landing And Following
Nitin J. Sanket,
Chahat Deep Singh,
Chethan Parameshwara,
Cornelia Fermüller,
Guido de Croon,
Yiannis Aloimonos
RSS, 2021
project page /
code /
arXiv /
video
We present a deep learning-based solution for detecting propellers (to detect drones) for mid-air landing and following.
|
|
|
EVDodgeNet: Deep Dynamic Obstacle Dodging with Event Cameras
Chethan Parameshwara*,
Nitin J. Sanket*,
Chahat Deep Singh,
Ashwin V. Kuruttukulam,
Cornelia Fermüller,
Davide Scaramuzza,
Yiannis Aloimonos
(* equal contribution)
ICRA, 2020
project page
/
code
/
arXiv
/
video
We present the first deep learning based solution for dodging multiple dynamic obstacles on a quadrotor with a single event camera and onboard computation.
|
|
Meta
Aug, 2025 -- Present
Researcher
Research on real-time diffusion models for personalized AI assistants.
|
|
WaveForms AI
May, 2025 -- Aug, 2025
Member of Technical Staff
Research on real-time multimodal diffusion models for audio-video generation.
|
|
Amazon
Aug, 2022 -- May, 2025
Researcher
Research on video diffusion models (text-to-video, image-to-video, video-to-video) and 3D computer vision (NeRF, Gaussian Splatting, Neural Rendering, and SLAM).
|
|
University of Maryland, College Park (UMD)
Aug, 2017 -- Aug, 2022
Graduate Research Assistant with Prof. Yiannis Aloimonos and Dr. Cornelia Fermüller
Research on video understanding (object detectiona and segmentation, motion segmentation) and 3D computer vision (pose estimation, SLAM) for autonomous vehicles.
|
|
SRI International (formerly Stanford Research Institute)
Summer, 2021
Research Intern with Dr. David Zhang and Michael Piacentino
Research on few-shot deep learning and gradient-free learning algorithms.
|
|
Reviewer: NeurIPS, ICLR, CVPR, ECCV, RAL, ICRA, IROS, IEEE Sensor Journal
NACS Representative, UMD Graduate Student Government (2020-2021)
Co-Chair, NACS Grant Review Committee (2019-2022)
Staff Member, Neuromorphic Cognition Engineering Workshop (July 2018)
|
Thanks to Jon Barron for this minimalist website template.
|
|