Computational Philology

Beyond Good and Evil,
Beyond Translation

Measuring how five English translations diverge from Nietzsche's German using sentence embeddings.

225Aphorisms Analyzed
6Versions Compared
0.806Max Fidelity

Explore the Translations

Compare how different translators render the same aphorism. Search for keywords, filter by divergence, or hit "Surprise me" for a random high-divergence passage.

Translator Fingerprints

UMAP projection of 1,356 embeddings reveals distinct clusters. Each translator leaves a stylistic signature the model can detect.

Hover to isolate clusters. German sits at center. Kaufmann and Hollingdale cluster nearby.

The Semantic Centroid

Pairwise cosine similarity reveals Hollingdale as the gravitational center—closest to the German (0.806) and closest to every other translator.

Similarity to Original German

The Full Matrix

Hover for exact scores. Kaufmann-Hollingdale (0.887) cluster tightest.

The Question

I kept switching translations—Kaufmann felt like a professor footnoting away the danger, Hollingdale felt rawer, Zimmern softened everything. Could NLP quantify what I felt?

"Was sich am schlechtesten aus einer Sprache in die andere übersetzen lässt, ist das tempo ihres Stils..."
— BGE §28

"That which translates worst from one language into another is the tempo of its style..."

Nietzsche argues that tempo is rooted in "the average tempo of its metabolism"— a language carries the physiological signature of its speakers. The translator who captures words but loses tempo has captured nothing.

Where They Diverge Most

Some passages translate consistently. Others scatter wildly—where the German underdetermines the English and translators must make choices no dictionary dictates.

§38σ = 0.323

French phrases, cultural critique

§130σ = 0.250

Talent reveals character

§74σ = 0.251

Genius requires gratitude

Short aphorisms diverge more (less context = more ambiguity). Passages with embedded French scatter wildly. Self-referential passages about language itself prove hardest to render consistently.

The Translators

Hollingdale, 1973

Self-taught translator, literary over academic. Sits at the semantic center— closest to German and everyone else. The faithful middle.

Kaufmann, 1966

Academic standard. Careful, scholarly, heavily footnoted. Clusters tightly with Hollingdale (0.887)—same era, same context.

Faber, 1998

Oxford World's Classics. Accuracy over style. Reliable but rarely surprising.

Norman, 2002

Cambridge edition. Takes interpretive risks, modernizes idioms. Drifts further from German but reads more easily.

Zimmern, 1906

First major translation. Victorian sensibilities filter the viciousness. Most distant from all others—the machine sees every year of that century.

Limitations

Sentence embeddings capture semantic similarity in web-text space, not philosophical fidelity. The model learned "meaning" from Wikipedia, not Zarathustra. A philosophy-tuned model was better at Nietzsche's concepts but worse at cross-lingual alignment. You cannot optimize for both.

What I measure: relative divergence patterns. Where translators cluster and scatter. Not which translation is "best"—but that interpretive schools exist and fingerprints are real.

The Orthography Problem

Nietzsche wrote before the 1901 spelling reform. The model sees "Werth" and "Wert" as different words. I built a 95-rule normalizer that improved alignment by 0.002-0.003. Small, but rankings stayed stable.

ArchaicModernSim.
giebtgibt~0.52
WerthWert~0.53
seynsein~0.48

Takeaway

Translators have fingerprints. Hollingdale sits at the semantic center. And §28, where Nietzsche writes about tempo and the untranslatable? One of the least divergent passages (σ = 0.024). Poetically fitting.