What was that again? A mathematical model of language incorporates the need for repetition

August 29, 2011, American Institute of Physics

As politicians know, repetition is often key to getting your message across. Now a former physicist studying linguistics at the Polish Academy of Sciences has taken this intuitive concept and incorporated it into a mathematical model of human communication.

In a paper in the AIP's journal , Łukasz Dębowski mathematically explores the idea that as humans we often repeat ourselves in an effort to get the story to stick. Using statistical observations about the frequency and patterns of word choice in natural language, Dębowski develops a model that shows repetitive patterns emerging in large chunks of speech.

Previous researchers have noted that long texts have more entropy, or uncertainty, than very brief statements. This tendency to higher entropy would seem to suggest that only through brevity could humans hope to build understanding – uttering short sentences that won't confuse listeners with too much information. But as long texts continue to get longer, the increase in the entropy starts to level off. Dębowski connects this power-law growth of entropy to a similar power-law growth in the number of distinct words used in a text. The two concepts – entropy and vocabulary size – can be related by the idea that humans describe a random world, but in a highly repetitive way.

Dębowski shows this by examining a block of text as a dynamic system that moves from randomness toward order through a series of repetitive steps. He theorizes that if a text describes a given number of independent facts in a repetitive way then it must contain at least the same number of distinct words that occur in a related repetitive fashion. What this reveals is that language may be viewed as a system that fights a natural increase in by slowly constructing a framework of repetitive words that enable humans to better grasp its meaning. For now the research is theoretical, but future work could experimentally test how closely it describes real texts, and maybe even candidates' stump speeches.

Explore further: Entropy study suggests Pictish symbols likely were part of a written language

More information: "Excess entropy in natural language: present state and perspectives" by Lukasz Dębowski is accepted for publication in Chaos: An Interdisciplinary Journal of Nonlinear Science.

Related Stories

Sign language puzzle solved

December 15, 2009

(PhysOrg.com) -- Scientists have known for 40 years that even though it takes longer to use sign language to sign individual words, sentences can be signed, on average, in the same time it takes to say them, but until now ...

Holographic dark information energy

May 30, 2011

Holographic Dark Information Energy gets my vote for the best mix of arcane theoretical concepts expressed in the shortest number of words – and just to keep it interesting, it’s mostly about entropy.

Linguists to re-think reason for short words

January 25, 2011

(PhysOrg.com) -- Linguists have thought for many years the length of words is related to the frequency of use, with short words used more often than long ones. Now researchers in the US have shown the length is more closely ...

Recommended for you

Crowds within crowd found to outperform 'wisdom of the crowd'

January 18, 2018

A team of researchers affiliated with institutions in Argentina, the U.S. and Germany has found that there is a way to improve on the "wisdom of the crowd"—separate the people in a given crowd into smaller groups and let ...

Study sheds new light on ancient human-turkey relationship

January 17, 2018

For the first time, research has uncovered the origins of the earliest domestic turkeys in ancient Mexico. The study also suggests turkeys weren't only prized for their meat—with demand for the birds soaring with the Mayans ...


Adjust slider to filter visible comments by rank

Display comments: newest first

not rated yet Sep 06, 2011
No. The meanings of words change with use.
Repetition can not prevent the change of the meanings of words repeated.
"Independent" facts are nonexistent.

What this reveals is that language may be viewed as a system that encourages "a natural increase in entropy" by slowly constructing a framework of non-replicable meaning for words.

A repeated word's meaning increases with use.
A new unique, and independent meaning is created each time the word is repeated.

What "moves from randomness toward order" is the AGREEMENT IN UNDERSTANDING that any NEW meaning a word acquires by use (repetition) is MEANINGLESS. To render any event or object psychologically, physically, biologically and chemically meaningless is impossible. So the will to ignore unavoidable and inevitable new meaning is invoke - stunting potential growth of understanding by agreement.

The human will to discard or accept this agreement determines the language's entropy. The choice between a static and dynamic system.
not rated yet Sep 06, 2011
New words are created to facilitate understanding.
Repeated words are used to limit understanding.

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.