publications

Aditya Kapoor, Yash Bhisikar, Benjamin Freed, Jan Peters, Mingfei Sun

Effective communication in Multi-Agent Reinforcement Learning (MARL) is bottlenecked by bandwidth constraints, yet standard approaches use inefficient fixed-precision vectors. We address this by developing a universal plug-and-play layer enabling agents to dynamically learn what to communicate and at what precision. Our method uses stochastic quantization with a reparameterization trick to pass gradients through discrete channels, employing a novel differentiable loss that bounds expected message length for unbounded signals. Agents naturally emerge with frequency-aware coding schemes—allocating fewer bits to common observations—effectively learning lossless compression akin to optimal coding theory. Across diverse benchmarks, our framework reduced bandwidth by one to five orders of magnitude—up to 45,000× in Traffic Junction—while boosting success rates by over 467% in complex tasks like Google Research Football. We demonstrate the "Bitter Lesson" in MARL: simple scalable architectures using general mechanisms outperform specialized communication architectures.

Under review at ICLR 2026

Mark Schöne†, Yash Bhisikar†, Karan Bania†, Khaleelulla Khan Nazeer, Christian Mayr, Anand Subramoney, David Kappel († equal contribution)

Traditional methods for processing sparse geometric data—such as point clouds from LiDAR sensors and event streams from neuromorphic cameras—either ignore the irregular spacing of coordinates or require computationally expensive pairwise operations (O(N²) complexity). By injecting the relative coordinate differences (Δₖ = tₖ - tₖ₋₁) directly into the Mamba SSM's step-size parameter, our model learns an exponentially-weighted interaction kernel that captures geometric structure. This formulation computes all N pairwise interactions in O(N) sequential steps or O(log N) parallel steps using the associative Scan primitive, avoiding both the quadratic complexity of transformers and the implicit spatial reasoning of prior SSM approaches. For event-based vision, it reaches 100% accuracy on all 11 classes of the DVS128-Gestures dataset—a first for any method—and 86.3% on Spiking Speech Commands. We also achieve roughly ~2% consistent improvements over the vanilla Mamba models on point-cloud benchmarks, demonstrating that explicit geometric encoding provides a powerful inductive bias across spatio-temporal sparse data modalities.

Accepted at NeVI Workshop @ ICCV 2025

Shreyans Jain, Yash Bhisikar, Surjya Ghosh, Timothy Pierson, Sougata Sen

Low-cost framework using single-antenna WiFi CSI measurements to monitor eating habits.

IEEE PerCom 2025 (Work-in-Progress) - Runner up for Best Paper

Harshvardhan Mestha, Tejas Agrawal, Karan Bania, Shreyas V, Yash Bhisikar

Reproducibility study improving CLIP's counting ability with lower computational resources on a smaller training subset.

Under review at ReScience C

Yash Bhisikar†, Nirmal Govindaraj†, Venkatavihan Devaki†, Ritu Anilkumar († equal contribution)

Compares genetic algorithms vs gradient-based optimizers for rainfall prediction over North-East India using U-Net architectures.

EGU General Assembly 2024

Arnav Borkar, Joel Tony, Hari Vamsi K. N, Tushar Barman, Yash Bhisikar, Sreenath T. M., Arnab K. Paul

Analyzes I/O performance of different workloads for various BeeGFS configurations, showing default settings lead to imbalanced data distribution.

IEEE CLUSTER Workshops 2023