Tuesday, April 25, 2023

A Loudspeaker-Compatible Photo-Phonology

Last weekend, I gave a talk at the 10th Language Creation Conference on creating languages that do not use the human voice, in which I went over four case studies of successively more-alien phonologies. (One of which I have previously blogged about here.) Israel Noletto called it a "must-watch" for any speculative fiction writers putting created languages in their stories! Turns out, I had extra time, and could've talked about a fifth... but when I put together my abstract, I thought I'd be hard-pressed to fit 4 case studies in half an hour, so I cut it out. And so, I shall now present case study #5 here, in blog form!

After noodling over the cephalopod-inspired phonology for a while (for context, go watch my talk), it occurred to me that human sign languages and cephalopod communication have in common the feature that you can't flood an area with a linguistic signal the way that you can with a disembodied voice from a speaker system--they have to be displayed on a screen with a certain defined spatial extent, and even if it's a very big screen, the components of the signal are still not evenly distributed throughout space.

So, could we create a light-based language that is broadcastable in the way that audio-encoded languages are? And what sort of creature could evolve to use such a system? Well, trivially, yes, we can--just encode the existing language of your choice in Morse code (or something equivalent), and pulse the lights in a room in the appropriate pattern. Heck, people actually do this sometimes (although more often in thriller movies than in real life). But designing a language whose native phonology is Morse code is just... not that interesting. It doesn't feel materially different from designing a language to use the Latin alphabet, for example. We need more constraints to spark creativity here! So, what else could we do to more directly exploit the medium of non-localized light? In Sai's terms, how could we design something that is natural to the medium?

A first thought is that light and sound are both wave phenomena, and one could just transpose sound waves directly into light waves, and use all the same kinds of tricks that audio languages do... except, it turns out that continuously modulating the frequency of light is considerably harder than modulating the frequency of sound. We can do it with frequency-modulated radio, but that's still not how we actually encode audio signals in radio, and similar technology just doesn't exist in the visible range. And if we look at how bioluminescence actually works in nature, no known organism has the ability to continuously modulate the frequency of their light output; they have a small number (usually just one) of biochemical reactions that produce a specific spectrum, and that's it.

But, a bioluminescent creature could do essentially the same thing we do with AM radio: ignore the inherent wave properties of the carrier signal entirely, and vary the amplitude over time to impose a secondary information-carrying waveform, which can be considerably more complex than the binary on/off of Morse signals, and can in fact have its own frequency and amplitude components. That doesn't mean high-contrast flashes couldn't still be involved--going back to nature again, the intraspecific visual signalling of fireflies, for example, is very Morse-like. But it can have more complex components, resulting in a higher bitrate that feels more suitable for a language that's on par with human languages in utility and convenience. Biological signal modulation can be done by controlling the rate of release of certain chemicals (e.g., the rate at which oxygen is introduced into a firefly's light organ to react with luciferin), or by physical motion of shutters to occlude the light to varying degrees (a common mechanism among, e.g., bioluminescent fish whose light is produced by symbiotic bacteria).

So, now we have a single-channel frequency-and-amplitude-modulable signal; the next obvious analogy to explore (at least obvious to me) is whistling registers (again, for context, go watch my talk, or listen to the Conlangery episode on Whistle Registers in which I talk about my conlang Tjugem). However, we can't directly copy whistling phonology into this new medium, precisely because we are ignoring the wave nature of the carrier signal; for a creature with a high visual flicker-fusion rate, perceivable modulation frequencies could be fairly high, but still nowhere near the rate of audio signals; rather, frequency information would have to occupy about the same timescale as amplitude information. In other words, varying "frequency" would give you a distinction between amplitude changes that are fast vs. slow, but it would be much harder to do things like a simultaneous frequency-and-amplitude sweep and keep each component distinguishable, the way you can with whistling. You could do it with flickering "eyelids" or chemical mixing sphincters (or, as Bioluminescent backlighting illuminates the complex visual signals of a social squid in the deep sea puts it, "by altering conditions within the photophores (41) or by manipulating the emitted light using other anatomical features")--trills in human languages introduce low-frequency components of about the right scale--but just as the majority of phonemic tokens in spoken languages are not trills, I would expect that kind of thing in a light-based language to be relatively rare. (Side note: perhaps audio trills and rapid light modulation could both be considered analogous to cephalopod chromatic shimmer patterns.)

So, the possibilities for a single-channel light-based phonology are not quite as rich as those for a whistling phonology, although the possibility of trilling/shimmering does help a bit (even though, AFAIK, no natural whistle register makes use of trilling). But, while the number of channels available to a given bioluminescent species will be fixed, the number of channels that we choose to provide when constructing a fictional intelligent bioluminescent creature is not! And if they have multiple light organs that allow transmitting on multiple different color channels simultaneously, then just two channels would allow them to exceed the combinatorial possibilities of human whistle registers.

Using this sort of medium for communication would have some interesting technological implications. Recording light over time is in some ways much more difficult than mechanically recording sound, but reproducing it is trivial. Light-based semaphore code systems for long-range communication with shuttered lanterns might be a blatantly obvious technology very early in history; and even if it cannot be mechanically recorded, if someone is willing to sit down for a while and manually cut out the right sequence of windows in a paper tape, mechanical reproduction of natural-looking speech could also occur at a very low tech level (especially if the language is monochromatic). Analog optical sound is in fact a technology that was really used in recent human history, and the reproduction step for a species using optical communication natively would be much simpler than it was for us, as there's no need for them to do the translation step from optical signal back into sound.

Now, there's a lot of literature on animal bioluminescence, but not a ton on specific signalling patterns used by different species... except for fireflies. So, if we want to move away from abstract theorizing and look at real-world analogs to extract a set of constraints for what a light-based language might look like, borrowing from firefly patterns is probably our best bet. Additionally, and in line with modelling off of fireflies, I am going to avoid using polychromatic signals, and see just see how far we can get with a single-channel design. After all, I already looked at a multi-channel / multi-formant signal system in the electroceptive phonology of Fysh A. I won't be sticking strictly to firefly patterns, because fireflies pretty much only use flashes, without significant variation in amplitudes, and that would end up being very Morse-like. However, per the US National Park Service, there are some interesting variations in the flashing patterns seen in various species; for example:
  • Long, low-amplitude glows (not really a flash at all).
  • Single, medium-amplitude flashes with long gaps.
  • Pairs of medium-amplitude flashes.
  • Trains of medium-amplitude flashes.
  • Single high-amplitude flashes ("flashbulbs").
I see a pattern going on here that may just be a coincidence, but seems like a plausible restriction on a bioluminescent alien: for a creature like a firefly which is using its own metabolic resources to produce light, rather than relying on symbiotic bacteria, there may be a maximum average rate at which power can be delivered to photophores, thus implying that, while you can glow at a low level indefinitely, brighter flashes, using more power all at once, entail a longer recovery period between flashes to "recharge". So, IF. YOU. ARE. SHOUTING. YOU. MUST. SPEAK. SLOWER. This is analogous to the amplitude-frequency dependence seen in the Fysh A electroceptive phonology.

So, let's go ahead and define three amplitude bands that phonemic segments might occupy, analogous to the frequency bands that organize whistling phonologies:

  1. A low band, which allows continuous glows and smooth waves.
  2. A middle band, where we have to pause between blinks, but we can blink fast enough for multiple blinks to constitute a single segment.
  3. A high band, where recharge pauses are too long for sequential blinks to be interpreted as a single segment.
These are sort of analogous to "places of articulation". Then, we can also define attack/decay characteristics for each blink--something like "manners of articulation":
  1. Slow attack vs. hard attack
  2. Slow decay vs. hard decay--only available in the low band; the upper bands only allow hard decay, since they use up all the luciferin!
And, furthermore, we can have a distinction between:
  1. "Tapped" -- a single amplitude peak.
  2. "Trilled" -- two or more close-spaced amplitude peaks (not available in the high band)
And, in the low band only, a unique distinction between short peaks and long peaks.

So, now we can map out the complete set of distinctive segments that might exist--the alien IPA!
  1. Low
    1. slow, short, slow, tapped
    2. slow, short, slow, trilled
    3. slow, short, hard, tapped
    4. slow, short, hard, trilled
    5. slow, long, slow, tapped
    6. slow, long, slow, trilled
    7. slow, long, hard, tapped
    8. slow, long, hard, trilled
    9. hard, short, slow, tapped
    10. hard, short, slow, trilled
    11. hard, short, hard, tapped
    12. hard, short, hard, trilled
    13. hard, long, slow, tapped
    14. hard, long, slow, trilled
    15. hard, long, hard, tapped
    16. hard, long, hard, trilled
  2. Mid
    1. slow, tapped
    2. slow, trilled
    3. hard, tapped
    4. hard, trilled
  3. High
    1. slow attack
    2. hard attack
And we could also have phonemic lengthening of the darkness following a hard decay for the tapped segments in the lower bands, which would give us an additional 10 possible segments, for a total of 32. Note that there's not really anything here that corresponds to "vowels". You might try to think of the low+long or low+slow-decay+trilled segments as vowels, or at least continuants, but they don't have the amplitude peaks that we would typically associate with human vowels as syllable nucleii. In fact, the whole basis of human syllable structure is missing! Instead, we might organize segments into larger units based on what kinds of segments can start or end those units--kind of like I did in Fysh A with initial and non-initial segments. The higher amplitude bands make it harder to follow up quickly with additional segments, so it would make sense if those are finals in larger, syllable-analogous units, and we end up with alien syllables that terminate in amplitude peaks rather than having them in the middle--kinda like all of their syllables are "CV" (but recall that we don't actually have a good analogy for vowels here!)

Now, with 32 different possible segments to choose from, with varying degrees of distinctiveness, not all languages in this phonetic space will use all of them, or choose exactly the same subset--just like human languages don't all use every possible human spoken phone! In particular, the low-band segments will be the most difficult to distinguish on average, due to being the "quiet"-est, so I would expect languages to vary significantly in exactly which low-band segments they utilize.

For purposes of this sketch, I'll select the following phonemes for maximal distinction:

  1. Low
    1. slow, short, slow, trilled - <w>
    2. slow, long, slow, tapped - <r>
    3. hard, long, slow, tapped - <t>
  2. Mid
    1. slow, tapped - <d>
    2. slow, trilled - <rr>
    3. hard, tapped - <k>
  3. High
    1. slow attack - <b>
    2. hard attack - <p>
Plus long <tt> and <dd>, exploiting the geminated-darkness feature, giving us a total of 10 distinct phonemes. As in the canine phonology sketch, that's not a ton (actually less than occur even in Rotokas, with its famously small phonemic inventory), but if we look at organizing the language in terms of possible syllables rather than possible segments, things look better. If we specify that every syllable must have a rise from a dark segment to a bright segment, and terminates with the brightest segment, as soon as we see a drop, then we get the following possible syllable types:

L>M: 16 possible syllables
L>H: 8 possible syllables
L>M>H: 32 possible syllables
M>H: 8 possible syllables

For a total of 64--and that's without allowing multiple segments of a single type per syllable! If we allow clusters of low or mid segments, we get multiplicative gains. Again, different languages of this same theoretical species could vary in what kinds of clusters they allow, just as, e.g., Russian differs from Hawai'ian, so perhaps there are small-phonology languages that allow no clusters, but for convenience let's say that in this sketch we'll allow either two low segments or two mid segments per syllable; then we get:

LL>M: 64 possible syllables
L>MM: 64 possible syllables
MM>H: 32 possible syllables
LL>M>H: 128 possible syllables
L>MM>H: 128 possible syllables

And suddenly, it has become impractical to write with a syllabary!

After my LCC presentation, I had a conversation with Biblaridion in which he pointed out an aspect of all of these non-IPA-codable languages that's directly relevant to writing stories with them: who can perceive them, and who can produce them? Audio and visually-coded languages like the canine sketch, the cephalopod sketch, and Tjugem can all be perceived by humans, so we could in principle develop receptive multilingualism in them, even if we couldn't produce them (and in the case of languages like Tjugem, we can even learn to produce them, even though they don't use typicaly human phonemes). This "firefly" phonology falls into that class as well--if humans can learn to decode morse code, surely we could learn to understand a firefly phonology, but we couldn't reply in the same language, or at least not in the same modality, without technological assistance. Fysh A presents a more extreme case--if there were, say, some intelligent star-nosed moles with electroceptive noses inhabiting the Fysh's world, they could gain receptive competence while being mute, but humans can neither produce nor even perceive the language without technological assistance. This suggests a new pathway for developing alien creatures: decide what communicative barriers you need in place to drive the plot, pick a modality that makes that work, and design your creatures to make it plausible for them to communicate in that modality. In fact, on further reflection, this seems to be exactly what H. Beam Piper did for Little Fuzzy (and you thought I would get through this whole post without an affiliate link! ha!)--the Fuzzies do communicate with sound, but in a frequency range that humans can neither hear nor replicate!

2 comments:

  1. As I watched your talk I began to wonder if you’ve ever read Krueger’s paper ‘Language and techniques of communication as theme or tool in sf’. Krueger harshly criticised sf writers for their unimaginative take on xenolinguistics. He’d be happy to listen to you, I believe. The paper was published back in the 60s. More recent literature, like China Miéville’s Embassytown (2011), has raised the bar in that regard. Still, in terms of otherworldly linguistics I’m yet to encounter anything like your ‘mad’ ideas :)
    I hope speculative writers get to see your paper and your blog. This way, academic writers like myself, who are interested in glossopoesis, will have much more to write about.

    ReplyDelete
    Replies
    1. I hadn't read it, but now I have! And it reminded me that I should add Cycle of Fire to my list of things to review, probably in comparison with Mission of Gravity and Close to Critical, other Hal Clement works which use the "they just learned our language" approach, and maybe Needle....

      Delete