EP50

EP50: Stochastic Composition — Xenakis and Probabilistic Music

Gaussian Pitch Clouds · Poisson Density · Exponential Duration · Rule 30 Cellular Automata
3:07 Probability TheoryStochastic ProcessesMusicologyComputational Music

Overview / 概述

In 1957, Iannis Xenakis premiered Pithoprakta — a work for 46 string instruments in which every note placement was derived not from melodic invention, but from probability distributions. No melody, no motivic development: only statistical clouds of pitch and density evolving in time.

中文: “不写音符,只写概率分布。一九五七年首演,一位建筑师用高斯分布、泊松过程和指数分布生成管弦乐——比现代 AI 音乐早了半个多世纪。他叫克塞纳基斯。”

This episode formalizes the three probability distributions at the core of Xenakis’s compositional system: a Gaussian distribution governing global pitch placement, a Poisson process governing local note density in time, and an exponential distribution governing note durations. Together, these three distributions form a complete probabilistic grammar for generating orchestral texture from mathematical first principles.

The episode also introduces cellular automata (specifically Rule 30) as a deterministic-but-chaotic alternative to purely random sampling, and compares Xenakis’s “cloud” model to the Markov chain model from EP21 (Markov Chains and AI Composition) — two radically different philosophies of stochastic music, producing radically different sonic results.

中文: “皮托普拉克塔的意思是’由概率决定的行为'。整部作品没有旋律线,只有音高密度的统计分布。四十六件弦乐器,每个人按各自的随机轨迹移动。”


Prerequisites / 前置知識


Definitions

Definition 50.1 (Gaussian (Normal) Distribution)

A random variable follows a Gaussian distribution with mean and variance , written , if its probability density function is

The distribution is symmetric about , with 68% of the probability mass within one standard deviation and 95% within two.

Worked Example (Pitch Distribution). In Pithoprakta, Xenakis set the global pitch register to be approximately where corresponds to the central register of the strings and determines the spread. If (MIDI note, middle of the string range) and semitones, then a pitch sampled as falls between 48 and 80 (four octaves) with probability 95%. Notes near C4 are most probable; notes near C2 or C6 are rare but possible.

Definition 50.2 (Poisson Process)

A Poisson process with rate parameter is a model for the arrival times of random events such that:

  1. The number of events in a time interval of length has a Poisson distribution:

  2. Events in disjoint time intervals are independent.

  3. The expected number of events in an interval of length is .

Worked Example (Note Density). If a string section is assigned rate notes per second, then the probability of exactly 3 notes occurring in a 1-second window is

A higher produces denser textures; Xenakis could modulate over time to create crescendos of density without specifying any individual note.

中文: “局部看:每个声部的音符密度服从泊松过程。”

Definition 50.3 (Exponential Distribution)

A random variable follows an exponential distribution with rate , written , if

The mean duration is and the distribution is memoryless: knowing that a note has already lasted seconds gives no information about how much longer it will last.

The exponential distribution arises naturally as the inter-arrival distribution of a Poisson process: if notes arrive at rate , then the waiting time between consecutive notes is .

Worked Example (Note Duration). If (on average 2 note-endings per second), then:

  • Mean duration seconds
  • : about 13% of notes last longer than 1 second

A small (slow rate) produces long, sustained tones — a sparse texture. A large produces rapid staccato events.

中文: “指数分布控制音符时值:平均时值越长,整体感觉越稀疏。这三者合在一起,构成了克塞纳基斯的概率音乐语法。”

Definition 50.4 (Elementary Cellular Automaton)

An elementary cellular automaton is a one-dimensional grid of cells, each in state , evolving in discrete time steps. At each step, every cell’s new state is determined solely by its current state and the states of its two immediate neighbors.

Formally, the update rule is a function . There are possible rules, indexed 0–255 by their binary encoding.

Rule 30 is defined by the lookup table:

Left, Center, Right 111 110 101 100 011 010 001 000
New state 0 0 0 1 1 1 1 0

The name “Rule 30” comes from reading the output row in binary: .

Musical Mapping. Starting from a single active cell in the center and iterating Rule 30 produces a triangular spacetime diagram. Each row represents one time step; the pattern of 1s and 0s encodes which pitch classes are simultaneously active at that moment — a deterministic but visually chaotic harmony sequence.

中文: “规则三十:看左邻、自身、右邻三个格子,查表得到新状态。迭代下去,产生看似随机但确定性的模式。”

Definition 50.5 (Monte Carlo Sampling)

Monte Carlo sampling refers to the technique of approximating a complex distribution or integral by drawing a large number of independent random samples from a known distribution.

In the composition context: given a probability density over pitch, time, or duration, draw independent samples . For large , the empirical histogram of samples approximates by the law of large numbers:

Xenakis’s UPIC system (Unité Polyagogique Informatique du CEMAMu) allowed composers to draw probability density curves on a graphical tablet; the computer then performed Monte Carlo sampling to generate note sequences from those drawn distributions.

中文: “克塞纳基斯的计算机程序 UPIC,允许作曲家在屏幕上画出概率密度函数,系统自动从中采样生成音符序列。”


Main Theorems / 主要定理

Theorem 50.1 (Law of Large Numbers — Convergence of Empirical Pitch Distribution)

Let be independent and identically distributed random variables with distribution (e.g., for pitch). Define the empirical mean

Then almost surely as . More generally, for any measurable set ,

Proof.

This is the strong law of large numbers (Kolmogorov, 1933). The decisive step is the observation that for i.i.d. square-integrable random variables, the Chebyshev inequality gives , which immediately implies convergence in probability. Strengthening to almost-sure convergence uses Borel-Cantelli: for any , the sum converges, so only finitely many deviations exceed almost surely.

Musical meaning: With 46 string instruments each independently sampling pitches from , the aggregate histogram of pitches converges to the Gaussian shape as the number of events grows. The composer specifies the shape of the texture, not the individual notes.

Theorem 50.2 (Memoryless Property of the Exponential Distribution)

Let . Then for all ,

The exponential distribution is the only continuous distribution with this memoryless property.

Proof.

Direct computation:

For uniqueness: suppose is a non-negative continuous random variable with the memoryless property for all . Let . The functional equation with continuous, non-increasing, and forces for some .

Musical meaning: A note currently being held has no “memory” of how long it has already lasted — the probability of it ending in the next instant is always , regardless of its age. This produces textures with no rhythmic expectation or metric pulse.

Theorem 50.3 (Superposition of Poisson Processes)

If independent instruments produce notes via Poisson processes with rates , then the combined note stream is also a Poisson process with rate

Proof.

For two independent Poisson processes and , their sum has generating function

which is the probability generating function of . The independence and stationarity properties of the individual processes carry over to the sum by construction. Induction extends the result to processes.

Musical meaning: 46 string players, each with their own Poisson rate, collectively produce a single Poisson process. The conductor (or the score) controls global texture density by specifying the sum , which can be distributed among players in many ways.

Prop 50.1 (Markov vs. Xenakis: Sequential vs. Cloud Structure)

The Markov chain model (

EP21

) and the Xenakis cloud model are fundamentally distinct in their dependence structure:

  • Markov chain: — each note depends on the immediately preceding note (first-order memory).
  • Xenakis cloud: All notes are drawn independently from a common distribution . No note affects any other.

These models are statistically distinguishable: the auto-correlation function decays exponentially to zero for a Markov chain (if the chain is ergodic), whereas for independent sampling it is zero for all exactly.

Proof.

For an ergodic Markov chain with transition matrix and stationary distribution , the -step auto-covariance satisfies . By the Perron-Frobenius theorem, geometrically, so the auto-covariance decays to zero at rate equal to the second-largest eigenvalue of .

For independent sampling, exactly for all by independence.


Numerical Examples

Example 1: Sampling three Gaussian pitches.

Let (MIDI, middle register), semitones. A standard normal sample gives pitch (A3). Sample gives (F♯4). Sample gives (F5). No pitch is more “correct” than another — each is a sample from the cloud.

Example 2: Poisson density comparison.

With notes/second, the probability of notes in a one-second window:

0
1
2
3
4
5+ remaining probability

With , the mode shifts to or , producing a much denser texture with near-zero probability of silence.

Example 3: Rule 30 from a single seed.

Initial row (15 cells): 000000010000000

After 4 iterations under Rule 30:

Row 0:  000000010000000
Row 1:  000000111000000
Row 2:  000001000100000
Row 3:  000011101110000
Row 4:  000100010001000

Each row’s pattern of 1s encodes which of 15 pitch classes sound simultaneously. The visual complexity grows rapidly from a single seed — a deterministic system producing apparently random output.

Example 4: Markov vs. Cloud auto-correlation.

A Markov chain on pitches with second-largest eigenvalue has auto-correlation decaying to zero with half-life steps. The Xenakis cloud produces for all , exactly. A statistician listening to the two pieces could, in principle, measure this difference from the score.


Musical Connection / 音乐联系

音乐联系

Xenakis’s Three-Distribution Grammar

In Pithoprakta (1955–56), Xenakis applied these three distributions in a layered hierarchy:

  • Global texture (minutes scale): the overall pitch center and spread of the Gaussian evolve slowly across the piece, creating large-scale arching contours.
  • Mid-level density (seconds scale): the Poisson rate is modulated — quiet passages have , climactic passages have or more.
  • Note level (milliseconds scale): individual note durations are exponentially distributed, so any single note’s length is unpredictable, though the average duration is prescribed.

This separation of structural levels — global shape, local density, individual events — anticipates modern multi-scale generative models by 60 years.

Contrast with Markov Composition (EP21)

Markov chain composition is a sequential model: the next note depends on the current note. This produces music with recognizable short-range patterns — if C is common after G, the music feels like it “wants” to return to C. The music has memory.

Xenakis’s cloud model is a population model: all notes are drawn independently from the same distribution. There is no memory, no expectation, no goal-directed motion. The result is not a melody with occasional surprises — it is a statistical cloud in which the melody concept is simply absent.

中文: “马尔可夫链是时序依赖的:下一个音符取决于当前状态,有记忆性。克塞纳基斯是整体统计的:所有音符同时从全局分布中独立采样,无记忆性。前者是顺序结构,后者是云状结构。同样的随机概念,产生截然不同的音乐逻辑。”

Connection to Diffusion Models

Xenakis’s insight — “do not write the result, write the distribution” — resonates with modern AI image and audio generation. Diffusion models (DDPM, 2020) also do not specify outputs directly; they specify a forward noise process and learn to reverse it, sampling from a learned distribution. The philosophical stance is identical: the creator designs the probability law, and the artwork emerges from sampling.

中文: “克塞纳基斯的思路与六十年后的扩散模型有惊人的类比——都是’不写结果,只写分布'——但两者是独立发展的,并非前驱与后继的因果关系。音乐生成的底层逻辑,在上世纪中叶就已经出现了。”

The key difference: Xenakis designed his distributions by hand based on mathematical and aesthetic judgment. Modern diffusion models learn their distributions from data. The endpoint — sampling music from a probability law — is the same.

UPIC and Interactive Probability Design

The UPIC system (1977) realized Xenakis’s vision computationally: composers drew curves on a graphics tablet, directly shaping probability densities over pitch and time. The computer performed Monte Carlo sampling from these hand-drawn distributions. This was arguably the first digital audio workstation based on probabilistic music theory rather than note-by-note sequencing.


Limits and Open Questions / 局限性与开放问题

  1. Independence assumption: Xenakis’s model treats all 46 string players as fully independent samplers. Real ensemble playing involves micro-synchronization, intonation adjustment, and room acoustics coupling — all of which introduce statistical dependence. Whether this dependence is musically significant or merely a performance artifact remains unstudied.

  2. Parameter selection: The model specifies how to sample once are chosen, but says nothing about how to choose these parameters. Xenakis used mathematical intuition and aesthetic judgment. A formal theory of parameter design — analogous to Schenkerian analysis for tonal music — does not exist for stochastic composition.

  3. Perception of statistical structure: Listeners cannot hear that a piece is “Gaussian” or “Poisson” in any direct sense. The perceptual correlates of changing (spread) or (density) are roughly “register width” and “busy-ness,” but the mapping from distribution parameters to perceptual attributes is poorly understood and likely highly context-dependent.

  4. Cellular automaton music: Rule 30 produces deterministic outputs — the same seed always gives the same music. This undermines the “randomness” ethos of stochastic composition. Moreover, the musical interest of Rule 30 harmonic sequences has never been formally evaluated against randomly sampled sequences or against human-composed music.

  5. Hybrid models: Modern neural sequence models (transformers) impose short-range dependencies (like Markov) and long-range dependencies (like global structure) simultaneously. Whether a Gaussian pitch cloud constrained by a neural language model is closer to Xenakis or to a Markov chain is an open conceptual question.

Conjecture (Perceptual Threshold for Poisson Texture Discrimination)

There exists a perceptual threshold such that listeners can reliably distinguish two Poisson textures with rates and at better than chance, but cannot distinguish from for .

Falsification criterion: A psychoacoustic experiment presenting listeners with paired tone clusters generated by Poisson processes at rates differing by , and measuring discrimination accuracy as a function of , would either find a well-defined threshold (confirming the conjecture) or find that discrimination accuracy varies continuously with with no clear threshold (falsifying the conjecture and suggesting a signal-detection rather than categorical model).


Academic References / 参考文献

  1. Xenakis, I. (1971). Formalized Music: Thought and Mathematics in Composition. Indiana University Press. (Revised edition: Pendragon Press, 1992.) — The primary source; contains Xenakis’s own mathematical exposition of stochastic composition, including the derivations underlying Pithoprakta.

  2. Xenakis, I. (1956). Pithoprakta for orchestra. Score published by Boosey & Hawkes. — Original orchestral work demonstrating the three-distribution grammar.

  3. Kolmogorov, A. N. (1933). Grundbegriffe der Wahrscheinlichkeitsrechnung. Springer. (English: Foundations of the Theory of Probability, Chelsea, 1956.) — The axiomatization of probability theory underlying the law of large numbers (Theorem 50.1).

  4. Feller, W. (1968). An Introduction to Probability Theory and Its Applications, Vol. 1, 3rd ed. Wiley. — Standard reference for Poisson processes, exponential distributions, and the memoryless property (Theorem 50.2).

  5. Wolfram, S. (1983). Statistical mechanics of cellular automata. Reviews of Modern Physics, 55(3), 601–644. — Systematic classification of the 256 elementary cellular automaton rules, including Rule 30.

  6. Wolfram, S. (2002). A New Kind of Science. Wolfram Media. Ch. 2 — Rule 30 as a source of complex behavior from simple deterministic rules.

  7. Pressing, J. (1988). Nonlinear maps as generators of musical design. Computer Music Journal, 12(2), 35–46. — Survey of deterministic chaos and cellular automata in music composition.

  8. Serra, X. (1997). Musical sound modeling with sinusoids plus noise. In Musical Signal Processing, eds. Roads et al. Swets & Zeitlinger. — Probabilistic noise models in music synthesis, connecting Xenakis’s distributions to modern audio synthesis.

  9. Ho, J., Jain, A., & Abbeel, P. (2020). Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33, 6840–6851. — The diffusion model paper; the “write the distribution, not the result” philosophy that parallels Xenakis.

  10. Solomonoff, R. (1964). A formal theory of inductive inference. Information and Control, 7(1), 1–22. — Formal basis for the relationship between algorithmic complexity and probability distributions, relevant to the question of what makes a “good” stochastic composition.

  11. Roads, C. (1996). The Computer Music Tutorial. MIT Press. Ch. 14 (Stochastic Composition). — Accessible survey of algorithmic and stochastic composition methods, situating Xenakis in the broader history.

  12. Hoffmann, P. (2009). The new GENDYN program. Computer Music Journal, 33(2), 31–32. — Update on Xenakis’s GENDYN stochastic synthesis system, the successor to UPIC.

  13. Solomos, M. (2021). From Music to Sound: The Emergence of Sound in 20th- and 21st-Century Music. Routledge. Ch. 3 (Xenakis). — Musicological context for Pithoprakta and Xenakis’s compositional philosophy.

  14. Norris, J. R. (1997). Markov Chains. Cambridge University Press. — Rigorous treatment of Markov chains, the contrasting model to Xenakis’s cloud; relevant to the comparison in Proposition 50.1.

  15. Moschovakis, Y. N. (1994). Notes on Set Theory. Springer. — For the measure-theoretic foundations underlying almost-sure convergence in Theorem 50.1.