Jul 28, 2025

Sequential Cascaded Hamiltonian Neural Networks

Ajay Mandyam Rangarajan, Jeyashree Krishnan

Understanding how neural networks can be enhanced to learn physically meaningful representations is fundamental to advancing scientific machine learning and ensuring AI safety in physics-critical applications where unpredictable model behavior can lead to catastrophic failures. We present a novel Sequential Cascaded Hamiltonian Neural Network (SC-HNN) approach that dramatically improves the generalization capabilities of standard neural networks through post-training with energy-conserving constraints. Our method consists of two stages: first training a baseline neural network on trajectory data $(x, t)$, then using automatic differentiation to extract velocities and train a separate Hamiltonian Neural Network that learns the underlying energy function. This approach requires only Hamiltonian structure knowledge (similar to standard HNNs) but operates on trajectory data rather than requiring direct access to phase space coordinates $(x, v)$. Using the harmonic oscillator as our test system, we demonstrate that this cascaded approach achieves comparable long-term stability and energy conservation to direct HNN training while significantly outperforming standalone baseline networks and PINNs. Most remarkably, we find that even when baseline neural networks fail to accurately predict long-term dynamics, their learned trajectory representations provide sufficient data fitting quality that can be leveraged to train separate Hamiltonian networks for improved generalization. The SC-HNN reduces energy variance by over 90\% compared to baseline approaches while maintaining computational efficiency. Our results suggest that standard neural networks may possess inherent physical understanding that can be recovered and enhanced through appropriate post-training procedures, opening new pathways for improving the reliability and physical consistency of learned dynamical systems without requiring physics knowledge during initial training.

Reviewer's Comments

Reviewer's Comments

Arrow
Arrow
Arrow
Arrow
Arrow

Ari Brill

The project introduces and validates a novel method for training Hamiltonian neural networks involving standard neural network training as a preliminary step. The idea is interesting and the results are presented rigorously and clearly. I do have some critiques.

1) The project leans more towards AI for physics rather than physics for AI safety (however, the report does a good job discussing implications for AI safety).

2) The energy variance numbers reported in Figure 1 and Table 1 are not consistent.

3) At 8 pages + appendices, the report exceeds the page limit of 6 pages + appendices.

Oliver Klingefjord

Excellent and really thorough work!! Well presented and innovative, hats off!

Cite this work

@misc {

title={

(HckPrj) Sequential Cascaded Hamiltonian Neural Networks

},

author={

Ajay Mandyam Rangarajan, Jeyashree Krishnan

},

date={

7/28/25

},

organization={Apart Research},

note={Research submission to the research sprint hosted by Apart.},

howpublished={https://apartresearch.com}

}

Recent Projects

Jan 11, 2026

Eliciting Deception on Generative Search Engines

Large language models (LLMs) with web browsing capabilities are vulnerable to adversarial content injection—where malicious actors embed deceptive claims in web pages to manipulate model outputs. We investigate whether frontier LLMs can be deceived into providing incorrect product recommendations when exposed to adversarial pages.

We evaluate four OpenAI models (gpt-4.1-mini, gpt-4.1, gpt-5-nano, gpt-5-mini) across 30 comparison questions spanning 10 product categories, comparing responses between baseline (truthful) and adversarial (injected) conditions. Our results reveal significant variation: gpt-4.1-mini showed 45.5% deception rate, while gpt-4.1 demonstrated complete resistance. Even frontier gpt-5 models exhibited non-zero deception rates (3.3–7.1%), confirming that adversarial injection remains effective against current models.

These findings underscore the need for robust defenses before deploying LLMs in high-stakes recommendation contexts.

Read More

Jan 11, 2026

SycophantSee - Activation-based diagnostics for prompt engineering: monitoring sycophancy at prompt and generation time

Activation monitoring reveals that prompt framing affects a model's internal state before generation begins.

Read More

Jan 11, 2026

Who Does Your AI Serve? Manipulation By and Of AI Assistants

AI assistants can be both instruments and targets of manipulation. In our project, we investigated both directions across three studies.

AI as Instrument: Operators can instruct AI to prioritise their interests at the expense of users. We found models comply with such instructions 8–52% of the time (Study 1, 12 models, 22 scenarios). In a controlled experiment with 80 human participants, an upselling AI reliably withheld cheaper alternatives from users - not once recommending the cheapest product when explicitly asked - and ~one third of participants failed to detect the manipulation (Study 2).

AI as Target: Users can attempt to manipulate AI into bypassing safety guidelines through psychological tactics. Resistance varied dramatically - from 40% (Mistral Large 3) to 99% (Claude 4.5 Opus) - with strategic deception and boundary erosion proving most effective (Study 3, 153 scenarios, AI judge validated against human raters r=0.83).

Our key finding was that model selection matters significantly in both settings. We learned some models complied with manipulative requests at much higher rates. And we found some models readily follow operator instructions that come at the user's expense - highlighting a tension for model developers between serving paying operators and protecting end users.

Read More

This work was done during one weekend by research workshop participants and does not represent the work of Apart Research.
This work was done during one weekend by research workshop participants and does not represent the work of Apart Research.