signatures of post-biological intelligence
TRANSCRIPT
Signatures of Post-Biological Intelligence: extending the search spectrum
Dr John R Elliott Reader in Intelligence Engineering
International Astronautics Association SETI Committee; Post-Detection Task Group
Leeds Metropolitan University, UK.
Random e.g White Noise
Highly predictable
e.g Pulsars
Structured Phenomena
Music
D.N.A
Long chain Polymers
All
Language...?
The Signal Universe
Would a machine construct [evolve] a communication system based on logic or an optimised form of natural language encoding [no redundancy - 100% pattern utilisation]?
We need not just target [listen to] areas of the Universe which support biological life.
We look at the likely signatures and contrasting structures such non-biological communicators may present and discuss how such contrasting forms of information exchange can aid, extend and refine our detection and decipherment capabilities.
To achieve this, we first look beneath the Veneer at known biological communicators and the common structures they demonstrate.
...do machines, or do post-biological agents expand the signal universe?
Surface Structure: all language [communication] demonstrates similar
structural surface entropic structure, as does information rich phenomena ;
Internal Structure: more specific to language – as increased dependency
is calculated, the entropic value diminishes at an approximate -1 slope [log-
log graph] ;
Cohesion: how sub-types of an information [communication] system
bonds with other sub-types, to build narrative;
Cognition: the constraints of a system that reflects the processing limits
of the users. This results in chunking of blocks of information, which
often share an imbedded relationship, within the narrative [discourse]:
e.g. Phrases.
The Human Language Template: some core features
The Efficient Compromise
Rhythm and structure – you may not understand the utterances but you can usually recognise what language is being spoken; Evolution of language – evidence and many simulation show that the efficient relationship of frequency and word length are a constant.
....linguistic reciprocal altruism
Human Language Evolution - a simulation
Over time the simulation follows similar paths for successive trials of grammar rule
induction. This begins with a large set of arbitrary rules when the size of the E-
language is small, which expand to an even larger set as the E-language expands,
before eventually ‘settling down’ to a concise and relatively small set of rules when
the E-language is extensive.
From such simulations, basic structural properties of language emerge over time
through the complex dynamical process of social transmission.
Human Language Evolution - a simulation
Grunts, babble, emergent structure.........refinement to efficiency, where
principles of least effort and reciprocal altruism are embodied.
Yep...me in the morning
Humans do it.....and so do Dolphins
Ultimately, we evolve:
• a very small number of very common words
• a small-medium number of middle frequency words
• a very large number of words that are infrequent
These simulations also demonstrate that frequently used words, such as
function words, eventually evolve to be the shortest.
It is of interest to note that while some dolphins
are reported to have learned English – up to 50
words used in correct context – no human being
has been reported to have leaned dolphinese.
Carl Sagan
Dolphins
Dolphin ‘Chatter’
2
5
3
7
162
108 137
20%
20%
20%
20%
85%
80%
13% 14%
27%
57% 13%
60%
67%
17%
• Number of distinctive sounds ≈163
A comparison of structure
= Dolphin
(Robot) Silicon chat Lingodroids are robots, which use an onboard camera, sonar and a laser range-finder, to map the space around them . The language, which sounds similar to the tones on a phone, is 'spoken' aloud by using a microphone and speaker. Experiments conducted in this project are a useful insight into how machine intelligence may develop communication.
These words have been invented by the robots themselves, using a variety of games to establish correlations between specific words and places, directions, and distances. And recently Lingodroids have been teaching themselves brand new words for different lengths of time.
Some examples of the Lingdroid vocabulary Geographical location examples: yifi, kiyi, gige, mira, xala, soqe, sihu, juhe, rije, pize, tuto, kopo, heto, qoze, yaro, zuce, xapo, zuya, fili, fexo, pucu, reya. Distance measurement examples: puga, puru, vupe. duka, ropi, puga, huzu, hiza, kobu, bula, Temporal examples: kafi, puni, fohu, qija, fedi, tofe
Unlike human discourse, robot [Lingodroid] vocalisation requires no inbuilt redundancy - like dialing phone numbers. So, optimisation of ngram usage - for a given word length - is both possible and desirable, for purposes of efficiency.
Robot Language Change, Learning and Sharing
Language Change ↔ Learning Period
When different communities of agents meet, their different
languages are acquired quickly: high frequency words are rapidly
well established.
Comprehension well established, even if terms not used by other
community.
Computer code, Robot & Human Entropic Profiles
0
1
2
3
4
5
6
7
1 2 3 4 5 6
Computer code
Lingodroid
Human Language
0
20
40
60
80
100
120
140
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48
LinCos vs English
Semantic chunks / phrases
Bigrams = 676 Trigrams = 17, 576 Tetragrams = 456,976 Pentagrams = 11, 881, 376 The above number of combinations is based on 26 letter alphabet, which is, of course, an assumption.
We use words that have a high relative frequency of use peaking at the range 8 to 10 letters. If we actually were able to utilise all combinations, this would provide us with a possible lexicon of 141,167,095,653,376 words. Or rounding down to 140 Trillion (USA) or 140 Billion in old UK currency!
Conclusions – for now
Entropic Signatures – similar initial profiles but differences do occur as higher entropic values are calculated; Redundancy and Ngram frequencies contrast significantly; No evidence (yet) for morphology in lingodroids;
The distribution of functional elements to logically relate content elements, in LinCos, are not constrained by cognition, as it is not a spoken conduit; Typical distributions found in natural language are currently absent in evidence available.
The signature of machine communication will be seen in the relationships and structure of the comprising patterns that represent semantic content, as it is clear that initial entropic profiling will be similar for both machine and biological intelligent communicators.
Conclusions – for now
So, whether Morse code, Ogham style encoding, Tonal encoding......or whatever conduit is used to encode and communicate information, it is the relationship of the patterns that is important, when ascertaining the type of author (man or machine): their constraints, behavioural categories, frequency - relative percentage – of combinations utilised and redundancy.