/sprites/<file> ADVERSARIAL VLA
robustness of vision-language-action systems under live adversaries
- ROBUST
- INTERP
- SCALE
- ADV
PHYSICAL · WORLD · AI · FIGHTING
/sprites/<file> /sprites/<file> Φ is an independent research collective for physical-world AI fighting. The symbol is Φ. We pronounce it "fight".
We study how physical-world AI systems — vision-language-action models, world models, and whatever architectural categories come next — learn, fail, and adapt when made to fight other physical-world AI systems. The fighting we care about is what happens inside the intelligence: the representations, the world models, the failure modes, the strategic emergence when one physical AI is set against another.
We are not a company. We are not a lab in any institution. We are a small group of independent researchers working at the intersection of multi-agent RL, VLA & world models, mechanistic interpretability, contact-rich dynamics, and adversarial self-play. Φ is a ten-year project. We publish openly.
active and planned research vectors
/sprites/<file> robustness of vision-language-action systems under live adversaries
/sprites/<file> world models when an opponent shapes the dynamics
/sprites/<file> mechanistic interpretability of self-play policies
/sprites/<file> strategic emergence under hard physical constraints
/sprites/<file> opponent-distribution coverage without league-scale compute
/sprites/<file> adversarial generalization across body morphologies
the people who do the work
FOUNDER · researcher
HKUST · undergraduate
Started Φ in 2026. Working on world models under adversarial dynamics and the mechanistic interpretability of self-play policies.
co-founder · pending
HKUST
Recruiting. If you are doing serious research in physical-world AI fighting and want a long-term home for it, the slot is open.
publications & status updates · open access by default
No preprints yet. First arXiv submissions expected late 2026 / early 2027. All Φ-affiliated work appears under open licenses with reproducible code, on arXiv and at appropriate venues (ICLR, NeurIPS, ICML, CoRL, RSS, and their workshops).
Mainstream physical AI optimizes single agents in cooperative or static environments. Pick up the cube. Walk on the terrain. Don't fall. This is not the regime in which physical intelligence will actually have to operate.
The hardest, most informative tests of a physical AI system are not how well it performs alone, but how it behaves when another physical AI is actively trying to defeat it. This is where:
The regime sits awkwardly between multi-agent RL (mostly toy domains), embodied AI (mostly single agents), and interpretability (mostly language models). Φ exists to occupy that intersection seriously.
serious researchers in adversarial physical AI, self-play interpretability, or contested world models — reach out.