Homepage
The Mathematics Colloquium in the Department of Mathematics & Computer Science at Rutgers-Newark takes place on Wednesdays 4-5pm, either in person at 204 Smith Hall, 101 Warren St., or via Zoom. All are welcome!
For more information or to be added to our mailing list, please email Kyle Hayden (kyle.hayden@rutgers.edu).
Schedule — Spring 2026
| Date | Speaker | Title |
|---|---|---|
| Jan 28 | Shira Wein (Amherst) | Lost in Translation, and Found: Detecting and Interpreting Translation Effects in Large Language Models |
| Feb 4 | Isaiah King (GWU) | Cyber Threat Hunting with Graph Deep Learning |
| Feb 11 | Bikash Kanungo | Learning Physics Across Scales: From Quantum Many-Body Theory to Atomistic Models |
| Feb 16 (*Monday 4pm*) | Soudeh Ghorbani (Johns Hopkins) | Unblocking AI: Understanding and Overcoming Datacenter Network Bottlenecks in Distributed AI |
| Feb 18 | Geyang Dai (Nat. U. Singapore) | Elliptic Chern Characters and Elliptic Atiyah–Witten Formula |
| Feb 25 | Tamás Darvas (Maryland) | TBD |
| Mar 2* | Erin Chambers (Notre Dame) | TBD (*part of the Distinguished Lectures in Topology, Geometry, and Physics) |
| Mar 4 | Dave Auckly (Kansas State University) | TBD |
| March 11 | David Futer (Temple) | TBD |
| March 18 | No colloquium | Spring break |
| March 25 | Heidi Goodson (CUNY) | TBD |
| April 1 | Hannah Turner (Stockton University) | TBD |
| April 8 | Sahana H. Balasubramanya (Lafayette) | TBD |
| April 15 | ||
| April 22 | ||
| April 29 | ||
| May 6 |
January 28
Shira Wein (Amherst)
Lost in Translation, and Found: Detecting and Interpreting Translation Effects in Large Language Models
Large language models are able to generate highly fluent text, in large part because they are trained on massive amounts of data. This data may contain “translationese”: hallmarks which distinguish translated texts from texts originating in the language. Though individual translated texts are often fluent and preserve meaning, at a large scale, the presence of translated texts in training data negatively impacts performance and in test data inflates evaluation. In this work, I investigate (1) whether humans are able to distinguish texts originally written in English from texts translated into English, (2) how the surface-level features of translationese can be mitigated using Abstract Meaning Representation, and (3) why neural classifiers are able to distinguish original and translated English texts much more accurately than humans.
February 4
Isaiah King (GWU)
Cyber Threat Hunting with Graph Deep Learning
Modern computer networks generate massive volumes of high-dimensional, time-evolving data, making the detection and response to cyber-attacks increasingly challenging. This challenge is especially acute for novel or zero-day attacks, where predefined signatures or heuristics are ineffective. This talk presents a framework for modeling large-scale computer networks as temporal graphs, enabling scalable, precise, and generalizable approaches to intrusion detection and incident response. Using deep graph deep learning techniques, including temporal link prediction and graph representation learning, anomalous activity can be identified in complex network environments. If an attacker is detected on the network, the same graph-based abstraction supports decision-making for active defense. Framing network defense as a multi-agent Markov game, graph-based reinforcement learning can be used to reason about adversarial behavior and select actions that contain and remove attackers while minimizing disruption to normal operations. This work lies at the intersection of cybersecurity and data science, advancing the state of the art in attack detection, attribution, and automated response to sophisticated adversaries operating in real-world networked systems.
February 11
Bikash Kanungo (Michigan)
Learning Physics Across Scales: From Quantum Many-Body Theory to Atomistic Models
Density functional theory (DFT) and atomistics have long remained the backbone of computational chemistry and materials science. DFT, being an electronic structure method, provides a quantum-mechanical description of interacting electrons. Atomistic methods, on the other hand, remove the electronic degrees of freedom to simulate the dynamics of atoms at length- and time-scales beyond DFT’s reach. Together, these approaches account for 40% of global scientific computing resources. Despite their success, both suffer from long-standing challenges in accuracy, transferability, scalability, and multiscale consistency. In DFT, the unknown exchange-correlation (XC) functional defines a nonlinear, nonlocal map from electron density to the energy. Current approximations to it remain far from chemical accuracy. Plus, the high computational demands of DFT limit their routine usage to a few thousand atoms. In atomistic modeling, interatomic potentials (IPs), be it classical or machine-learned, are typically fit to narrow datasets, limiting their transferability. More importantly, the lack of electronic information in IPs limits their ability to describe electronically driven phenomena, such as in surface chemistry, emergent behavior in quantum materials, and biochemical reactions.
In this talk, I will present various approaches to address these fundamental challenges in DFT and atomistics. First, I will show how accurate quantum many-body data can be used to machine-learn the XC functionals by solving the inverse DFT problem, yielding systematically improvable and physically constrained models. Second, I will introduce a new approach, termed field-theoretic atomistics (FTA), as a large-scale machine-learned surrogate for DFT that adheres to the known physical symmetries and variational principles. Unlike IPs, FTA retains electronic degrees of freedom at similar computational costs as current machine-learned IPs. Finally, I will also discuss how, combined with fast and scalable numerical methods, this approach of integrating machine learning with physical principles can overcome long-standing accuracy and scale barriers in quantum mechanical modeling of materials.
February 16 (Monday)
Soudeh Ghorbani (Johns Hopkins)
Unblocking AI: Understanding and Overcoming Datacenter Network Bottlenecks in Distributed AI
As companies continue to invest heavily in AI-dedicated datacenters, a critical yet often underestimated challenge persists: datacenter networks remain a major bottleneck in distributed AI training. Despite rapid advances in compute hardware and machine learning algorithms, network congestion and communication overhead still limit the scalability and efficiency of large-scale AI workloads.
In this talk, I will present insights from a comprehensive study I led, where our team instrumented and analyzed traffic across 20+ AI datacenters of a major hyperscaler. Our investigation revealed key characteristics of AI workloads, the root causes of persistent network bottlenecks, and the challenges that arise when attempting to mitigate them. Building on these findings, I will introduce new datacenter network designs that challenge long-standing paradigms, such as strict shortest-path routing and in-order packet delivery, by embracing more flexible, robust strategies. I will show how these approaches pinpoint and alleviate bottlenecks, yielding substantial performance improvements. I will conclude with open research questions and future directions in optimizing networks for AI at scale.
February 18
Geyang Dai (National University of Singapore)
Elliptic Chern Characters and Elliptic Atiyah–Witten Formula
A principal G-bundle over a manifold X, equipped with a connection, together with a positive-energy representation, gives rise to a circle-equivariant gerbe module on the free loop space LX. From this data we construct an elliptic Chern character on LX, and a refinement, the elliptic Bismut–Chern character on the double loop space.
We also generalize the Atiyah–Witten formula to double loop space. We show that the four Pfaffian sections, corresponding to the four spin structures on an elliptic curve, are identified with the four elliptic holonomies arising from the four virtual level one positive-energy representations when G=Spin. These constructions are closely related to conformal blocks in Chern–Simons gauge theory.