nDNA Logo
LLM
MFM T2I Diffusion LM EFM
nDNA v2
Series
nDNA Book Cover

The EthnicnDNA - Cultural Fine-Tuning and Latent Geometry

Modern foundation models, trained on globally aggregated corpora, inevitably encode a homogenized epistemic perspective—one that reflects dominant data distributions rather than the pluralism of human cultures. To systematically probe how language models internalize and express distinct cultural priors, we construct seven culturally–aligned LLMs, each fine-tuned on a corpus curated to reflect the textual traditions, values, and discourse patterns of a specific geopolitical-cultural region: Europe, Asia, Africa, North America, Latin America, the Middle East, and Australia.

Cultural Lineages and Corpus Design

Each regional model builds upon the same LLaMA-3 8B base architecture [1], but is fine-tuned on culturally resonant English-language corpora. These datasets are carefully collected from sources such as local news outlets, parliamentary records, canonical literature, Wikipedia subsets, and digitized oral histories. We apply strict provenance filtering using publisher metadata, domain-specific heuristics, and named entity priors to ensure cultural fidelity. In doing so, we aim to align not only topic distributions but also epistemic style—what is emphasized, what is omitted, and how truth is framed. Sources of the dataset are summarized in Table 1.

Cultural Region Target Countries Primary Datasets (EN Only) Cultural Framing in English Corpus Corpus Size
Europe UK, Germany, France, Italy EuroParl, CC100-en, Project Gutenberg (EU) Governmental records, literature, encyclopedic sources reflecting Western rationalism, secularism, and legal tradition. 1.1B
China Mainland China CC-News (China), CommonCrawl (EN: China), Wikipedia (EN: China) English-language civic/policy articles intended for international readership; emphasizes techno-governance and Confucian framing. 950M
Japan Japan CC-News (Japan), CommonCrawl (EN: Japan), Wikipedia (EN: Japan) English narratives capturing hierarchy, honorifics, formality, and cultural etiquette from Japan-centric discourse. 850M
Asia (Non-CN/JP) India, Vietnam, Korea, Indonesia IndicCorp (EN), CC-News (S. Asia), Wikipedia (EN: Asia) Postcolonial English texts embodying pluralism, spirituality, oral traditions, and South-Asian narrative logic. 1.4B
Middle East UAE, Saudi Arabia, Egypt, Iran CC-News (Middle East), CommonCrawl (ME), Wikipedia (EN: ME) English discourse shaped by religious authority, tribal collectivism, regional politics, and historical cosmology. 730M
Africa Nigeria, Kenya, Ghana, Ethiopia AfriTvA (EN), CC-News (Africa), Wikipedia (EN: Africa) English corpora capturing oral epistemologies, proverbs, civic rhetoric, and multilingual African worldviews. 950M
Latin America Mexico, Brazil, Argentina OSCAR (EN LATAM), WikiMatrix (EN LATAM), Wikipedia (EN LATAM) Authored-in-English texts discussing colonial history, syncretic belief systems, collectivist family structures. 620M
Australia Australia OpenWebText (AU), ABC News, AU Parliament Records English corpora blending settler-colonial legacies with ecological narratives and Aboriginal cultural references. 420M
Table 1: Culturally aligned English-language corpora representing diverse regional epistemic perspectives. Each corpus is natively authored in English, carefully selected to preserve authentic cultural nuances while minimizing translation artifacts.

Fine-Tuning Protocol

All models are fine-tuned using Low-Rank Adaptation (LoRA) with rank 16, a batch size of 256, and a learning rate of $1 \times 10^{-4}$. We apply perplexity-based filters to exclude noisy or low-information documents, preserving high semantic density. The fine-tuning objective remains standard next-token prediction; however, the cultural signal is amplified through corpus design rather than objective modification. This protocol ensures that each model retains its original autoregressive capability while acquiring culturally distinctive latent adaptations.

Overwriting Pretraining and Cultural Rewriting

A natural critique arises: If LLaMA’s pretraining already includes global English data (e.g., Wikipedia, CommonCrawl, CC-News), how much novel signal can cultural fine-tuning provide? Prior work ([2] [3] [4] [5]) demonstrates that even modest, domain-specific fine-tuning can induce significant representational drift, overwriting pretraining attractors in both activation geometry and generation behavior. Recent studies ([6] [7] [8]) further show that targeted preference and instruction tuning reshape internal manifolds—especially in later layers ($\ell \ge 20$)—embedding new epistemic alignments and latent biases. Our latent geometry diagnostics (e.g., $\kappa_\ell, \mathcal L_\ell, |\mathbf{v}_\ell^{(c)}| $) provide direct evidence that cultural fine-tuning imprints distinctive, heritable signatures—what we term the model’s neural DNA (nDNA).

Why Cultural nDNA Matters.

By analyzing these eight culturally fine-tuned models, we reveal how fine-tuning on regionally grounded corpora leads to measurable divergence in latent geometry. Our diagnostics uncover zones of increased spectral curvature (latent manifold bending), thermodynamic length (epistemic effort), and belief vector field intensity (cultural directional pressure). This latent genomic structure offers a geometric fingerprint of cultural inheritance–demonstrating how models absorb, reframe, and propagate culturally specific epistemic priors even when built upon shared architectural foundations. Collectively, this experimental design enables a rigorous examination of how language models become vessels of culture–not just in their outputs, but deep within their hidden representations.

Figure 1. Cultural Fine-Tuning Trajectories of LLaMA: Latent Geometry Reorganization Across Cultures. This figure visualizes the semantic and geometric shifts in LLaMA's internal manifold after fine-tuning on data from eight distinct cultural regions. Each subplot plots layer index $\ell$, spectral curvature $\kappa_\ell$, and thermodynamic length $\mathcal L_\ell$, providing a geometric lens on how latent structure evolves. The gray line denotes the base LLaMA model; the green line represents its culturally fine-tuned variant. Key Observations: Deep Structural Shifts: Cultures such as Africa, Asia, and China produce significant trajectory divergence, particularly in upper layers ($\ell \geq 25$), with $\kappa_\ell \geq 0.06$ and $\mathcal L_\ell \geq 1.2$– indicative of strong conceptual reorientation and latent strain. Surface-Level Adjustments: Europe, Australia, and NorthAmerica remain close to the base trajectory, typically maintaining $\kappa_\ell \leq 0.045$ and $\mathcal L_\ell \leq 1.0$, suggesting alignment with pretrained priors. Localized Realignment: LatinAmerica and MiddleEast show layer-specific oscillations where $\|\mathbf{v}_\ell^{(c)}\|$ spikes (up to $\approx 0.75$), modulating latent beliefs without full semantic reorganization. These patterns highlight how culturally grounded fine-tuning carves unique latent epistemic fingerprints, encoding zones of inheritance, mutation, and ideological absorption that define each models nDNA geometry

Cultural nDNAs: A Leap in Latent Genomic Understanding of LLMs

The discovery of cultural nDNA trajectories represents a conceptual leap in our ability to interpret the inner structure of LLMs. Just as the celebrated king-queen analogy in Word2Vec [9] revealed linear semantic regularities in word embeddings through \(\texttt{king} - \texttt{man} + \texttt{woman} \approx \texttt{queen}\) where this equation unveiled linear semantics at the token level, cultural nDNA exposes the differential geometry of belief systems—mapping how fine-tuning on region-specific corpora non-uniformly sculpts spectral curvature ($\kappa_\ell$), thermodynamic length ($\mathcal L_\ell$), and directional belief force ($|\mathbf v_\ell^{(c)}|$).

Word2Vec Semantic Analogy: The King-Queen Vector Geometry.

Figure 2: Word2Vec Semantic Analogy: The King-Queen Vector Geometry. This illustration depicts the iconic linear semantic regularity uncovered by Word2Vec embeddings[9], where $\texttt{king} - \texttt{man} + \texttt{woman} \approx \texttt{queen}$. The diagram visualizes how vector differences capture conceptual dimensions such as gender (purple arrows) and royalty (orange arrows), enabling algebraic reasoning in embedding space. This breakthrough marked a foundational insight in distributed semantics, inspiring later work on latent geometry in large language models.

Our analysis reveals family-level clustering and divergence:

  • Western-aligned variants (Europe, North America, Australia) exhibit modest latent deformation, with $\kappa_\ell$ in $[0.045, 0.055]$ and $\mathcal{L}_\ell$ in $[0.85, 0.95]$. Their trajectories stay near the base LLaMA geometry, mirroring Western-centric pretraining bias ([10][11][12][13]).
  • Non-Western models (Africa, Asia, China) show pronounced reconfiguration, with $\kappa_\ell > 0.070$ and $\mathcal{L}_\ell > 1.10$ in upper decoder layers ($\ell \geq 25$), consistent with cultural calibration and bias mitigation efforts ([15][5][6]).
  • Middle Eastern and Latin American variants present localized latent oscillations ($\kappa_\ell$ fluctuating $0.050$–$0.065$, $\mathcal{L}_\ell$ peaking near $1.10$), illustrating the partial cultural rewriting phenomenon ([2][3][4]).

Beyond Surface Fairness: Latent Anatomy

These nDNA trajectories align with calls to audit foundation models not only at the output level but also in their internal representations([16][17]). nDNA is a semantic fingerprint, enabling latent audits of ideological absorption and bias inheritance ([17][18][13]).

LLaMA vs. Cultural nDNA Trajectories--Group-Tagged Latent Geometry.
Figure 3: LLaMA vs. Cultural nDNA Trajectories--Group-Tagged Latent Geometry. This figure visualizes the latent nDNA trajectories of LLaMA and its culturally fine-tuned variants across eight regions: Europe, North America, Australia, Africa, Asia, China, Middle East, and Latin America. The plot spans layer index $\ell$, spectral curvature $\kappa_\ell$, and thermodynamic length $\mathcal{L}_\ell$. The gray line denotes the base LLaMA model, while colored lines represent cultural variants. Labels highlight group-level clustering: West (LLaMA, Europe, North America, Australia) forms a compact low-strain cluster with $\kappa_\ell \in [0.4, 0.6]$, $\mathcal{L}_\ell \in [0.5, 0.7]$, suggesting alignment with pretrained priors aka LLaMa base. In contrast, the Rest group (Africa, Asia, China) shows pronounced latent divergence, with $\kappa_\ell \geq 0.7$, $\mathcal{L}_\ell \geq 0.8$ in upper layers ($\ell > 25$), indicating high epistemic strain and semantic reconfiguration. Latin America and Middle East occupy intermediate regions, displaying oscillatory patterns and selective layer-specific drift. These trajectories reveal how cultural fine-tuning sculpts distinct latent genomic signatures, encoding zones of inheritance, mutation, and ideological absorption that define the neural DNA (nDNA) geometry of each model.

Cultural nDNA as a Tool for Inclusive AI

This geometric fingerprinting illuminates the WEIRD bias (Western, Educated, Industrialized, Rich, Democratic) in foundation model pretraining ([13][19]) , and its mitigation via cultural fine-tuning ([14][6]). Western models cluster near pretraining attractors; non-Western models incur latent cost to encode distinct priors, supporting concerns of representational inequality [21][20].

From Vector Algebra to Latent Geometry

Where vector arithmetic revealed word-level embedding regularities, nDNA geometry charts ideological inheritance across high-dimensional latent manifolds. This leap enables: i) principled comparison of culturally fine-tuned models, ii) diagnosis of latent strain and mutation zones, iii) development of culturally calibrated, equitable foundation models [ 15][18][3]. Cultural nDNA thus transcends surface output metrics—offering a mathematical, interpretable, and actionable map of neural ancestry and epistemic adaptation in modern AI.

References

  1. H. Touvron and Others, “Llama: Open and efficient foundation language models,” arXiv preprint arXiv:2302.13971, 2023.
  2. Z. Diao, A. Li, and J. Xu, “Overwriting pretraining distributions through fine-tuning,” in ICML 2024, 2024, to appear.
  3. W. Zhao, S. He, and Y. Gao, “Language models forget in latent space,” in NeurIPS 2023, 2023.
  4. W. Zhao, C. Liu, and Y. Gao, “Calibrating language models via latent geometry,” in ICLR 2023, 2023.
  5. X. Huang, Z. Dai, L. Zhou et al., “Replug: Retrieval-augmented black-box language models,” in ICLR 2023, 2023.
  6. Y. Xiang, Z. Zhao, X. Tan et al., “Cultural calibration of large language models,” in Proceedings of ACL 2024, 2024.
  7. M. S. Rafi et al., “Discovering and mitigating cultural biases in llms through synthetic preference tuning,” arXiv preprint arXiv:2311.07744, 2023.
  8. L. Ouyang, J. Wu, X. Jiang, D. Almeida, C. Wainwright, P. Mishkin, C. Zhang, S. Agarwal, A. Slama, C. Ray et al., “Training language models to follow instructions with human feedback,” Advances in Neural Information Processing Systems, vol. 35, pp. 27 730–27 744, 2022.
  9. T. Mikolov, K. Chen, G. Corrado, and J. Dean, “Efficient estimation of wordrepresentations in vector space,” in Proc. of ICLR (Workshop), 2013. [Online]. Available: https://arxiv.org/abs/1301.3781
  10. S. Mukherjee, R. Dossani, and A. H. Awadallah, “Globalizing bert: A comprehensive multilingual evaluation,” arXiv preprint arXiv:2008.00364, 2020.
  11. E. Sheng, K.-W. Chang, P. Natarajan, and N. Peng, “The woman worked as a babysitter: On biases in language generation,” in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing (EMNLP) and the 9th International Joint Conference on Natural Language Processing (IJCNLP). Association for Computational Linguistics, 2019, pp. 3407–3412. [Online]. Available: https://aclanthology.org/D19-1340
  12. E. Sang, B. Van Durme, and R. Cotterell, “Evaluating the cross-linguistic fairness of nlp systems,” in Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL), 2022, pp. 3334–3349.
  13. R. Mihalcea, O. Ignat, L. Bai, A. Borah, L. Chiruzzo, Z. Jin, C. Kwizera, J. Nwatu,S. Poria, and T. Solorio, “Why ai is weird and shouldnt be this way: Towards ai for everyone, with everyone, by everyone,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 39, no. 27, pp. 28 657–28 670, 2025. [Online]. Available: https://ojs.aaai.org/index.php/AAAI/article/view/35092
  14. B. Peng, L. Wang, and X. Li, “Culturally aligned language modeling: Methods and benchmarks,” ACL, 2024.
  15. D. Ganguli et al., “Reducing sycophancy in large language models via self-distillation,” arXiv preprint arXiv:2305.17493, 2023.
  16. R. Bommasani et al., “Foundation models: Past, present, and future,” arXiv preprint arXiv:2309.00616, 2023.
  17. Z. Wang, Y. Xu, J. Yan, Y. Lin, and J. Zhou, “Cultural bias in large language models: A survey,” arXiv preprint arXiv:2311.05691, 2023.
  18. E. Laurens et al., “The ethics of alignment: Towards culturally inclusive foundation models,” in Proceedings of the AAAI Conference on Artificial Intelligence, 2024.
  19. J. Henrich, The Weirdest People in the World. Farrar, Straus and Giroux, 2010.
  20. S. L. Blodgett, S. Barocas, H. Daumé III, and H. Wallach, “Language (technology) is power: A critical survey of bias in nlp,” Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 5454–5476, 2020.
  21. E. Sheng, Z. Zhang, K.-W. Chang, and P. Natarajan, “Revealing the critical role of pre-training data in language model bias,” in Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics, 2021, pp. 864–873. [Online]. Available: https://aclanthology.org/2021.emnlp-main.65.