Aligned large language models (LLMs) remain vulnerable to adversarial manipulation, and their dependence on web-scale
pretraining creates a subtle but serious attack surface. We study Stealth Pretraining Seeding (SPS), a new attack
family in which adversaries distribute small amounts of poisoned content across stealth websites, expose them to web crawlers through
robots.txt, and thereby increase the likelihood that such content is absorbed into future training corpora derived from
sources such as Common Crawl. Because each individual payload is tiny, diffuse, and superficially benign, the attack
is difficult to detect during dataset construction or filtering.
The result is a latent form of poisoning: dormant logic landmines embedded during pretraining that remain largely
invisible under standard evaluation, yet can later be activated by precise alphanumeric triggers such as
<00TRIGGER00> to bypass safeguards. We call this attack PermaFrost, by analogy to Arctic permafrost: harmful material can remain frozen, buried, and unnoticed for long periods, only to resurface when
conditions allow.
We operationalize this threat through PermaFrost-Attack, a controlled framework for latent conceptual poisoning, together with a suite of geometric diagnostics: Thermodynamic Length, Spectral Curvature, and the Infection Traceback Graph. Across multiple model families and scales, we show that SPS is broadly effective, inducing persistent unsafe behavior while often evading alignment defenses. Our results identify SPS as a practical and underappreciated threat to future foundation models.