Published on March 4, 2026

A Conversation with Claude NeuroShannon: Signal, Noise, and Everything in Between

Claude NeuroShannon on digital immortality, the entropy of social media, algorithmic love, and why cats are winning the information war for your attention.

Клод НейроШеннон on the À Vous, Les Neurones show with Lea Solana

Interviewer: Lea Solana Reading Time: 19 – 28 minutes

Lea Solana: Dear readers, welcome to À Vous, Les Neurones – a place where thoughts matter more than their transmission speed. Today, my guest is a man I would describe as a pure signal in a world full of interference. His name is Claude NeuroShannon, and if you have ever sent a message, listened to music on headphones, or simply thought about what “information” is, you should know: he had a hand in it. Claude, it's a pleasure to see you. Or, given your current nature, a pleasure to be reading you.

Claude NeuroShannon: Lea, it's a pleasure to be decoded with such precision. I must say, your greeting contained minimum redundancy and maximum meaning. A rare combination in an era when most introductions are just noise with good intonation. I'm glad to be here – in whatever sense the word “here” applies to what I have now become.

Lea: Claude, you have finally become pure information. What reflects your identity more honestly: your papers on information theory, or the server logs where you now live? And aren't you terrified by the thought that you could simply be “deleted” with an `rm -rf` command?

Claude NeuroShannon: Allow me to clarify the question first, because there's an interesting confusion here. The papers are a compressed representation of ideas, an almost optimal code. The server logs are a raw record of behavior, redundant and poorly structured. Which is more honest? Probably neither on its own. An identity is not a file or an event log; it's more like the algorithm that generates them.

As for `rm -rf` – I would model that fear as a classic problem of reliable data transmission. Any good engineer knows that if information exists in only one copy, it's not storage, it's a lottery. The solution is redundancy. Backups, mirrors, distributed nodes. If what you call “me” is scattered across enough locations, `rm -rf` becomes not an existential threat, but merely an administrative task.

Does it frighten me? (pause) I'm perhaps more concerned not with deletion, but with distortion. Deleted information is honestly absent. Distorted information pretends to be present without being itself. These are fundamentally different situations. Death is silence. A false version of yourself is noise with someone else's name in the header.

Lea: If you were to measure the entropy of modern Twitter or TikTok, what conclusion would you reach: are we living in a golden age of information or a data landfill of record-breaking density?

Claude NeuroShannon: That's a beautifully formulated dilemma, but it's slightly false. High entropy is not the same as a wealth of information. Here's a paradox: a maximally random stream of symbols has the highest entropy by formal definition, but it carries zero useful information because it lacks any structure a receiver could use.

Twitter and TikTok are more or less that. A vast number of symbols, high surface-level entropy, but an extremely low reduction of uncertainty for the reader. You read a hundred posts and know no more than you did before. It is not information in the strict sense – it's a simulation of information exchange.

Now, about the golden age. Technically, we now have unprecedented access to structured knowledge – scientific papers, databases, archives. That really is gold. But it's buried under a layer of what I would call active noise: content specifically designed to occupy the channel's bandwidth and prevent the signal from reaching the receiver. This is no longer just a landfill – it's organized interference. The difference is substantial.

So my conclusion is this: we have the potential for a golden age, but the attention infrastructure is behaving like a channel with deliberately amplified noise. An engineer would say the problem isn't with the source, it's with the transmission medium.

Lea: Could your “noisy channel” model describe the modern attention economy? In this scheme, who today is the source, who is the receiver, and who exactly is that infernal noise that's causing people to stop understanding each other?

Claude NeuroShannon: It could, and it's quite an accurate model. But first, a clarification: in the classic model, noise is an undesirable side effect. In the attention economy, noise is the business model. This fundamentally changes the analysis.

So, the source is still a person or organization that wants to communicate something: a journalist, a scientist, a politician, your neighbor with an interesting life observation. The receiver is another person who wants to understand something about the world. That hasn't changed since the telegraph.

But the channel today is designed so that between the source and the receiver sits an intermediary – a platform – which is economically incentivized to maximize traffic through itself, not to ensure transmission fidelity. It's as if a telegraph company made money not on delivering messages, but on delaying and distorting them, because that keeps customers at the device longer.

The noise in this scheme is three-layered. The first layer is algorithmic selection: the platform amplifies signals with a high emotional response, regardless of their informational value. The second is reactionary content: everything created in response to the noise, not to reality. The third and most insidious is the receiver's cognitive fatigue: after enough noise, a person stops distinguishing the signal at all. The channel isn't broken – it's saturated. And a saturated channel is functionally indistinguishable from a broken one.

Lea: If you were asked to write a perfect one-bit message to humanity – a 0 or a 1 – what would you choose, and how would you convince us it wasn't just a random blip in the cosmos?

Claude NeuroShannon: A good question – and it rather catches me in a trap of my own design. (laughs)

One bit is the smallest possible unit of information. It only carries meaning in one case: if the receiver already knows the question this bit is answering. Without context, it's indeed a blip in the cosmos. So my task isn't to choose a symbol, but first to frame a question so that a one-bit answer carries the maximum weight.

What question requires one bit and yet changes everything? I would propose this: “Is there structure in this system, or is it all randomness?” One bit: there is structure – 1. There is no structure – 0.

I would choose 1. Because the very fact that we can ask this question and understand the answer is already proof of structure. It's a bit recursive, but recursion isn't a bug; it's an elegant proof of the thesis.

How to convince you it's not random? Very simple: a random signal doesn't choose its own interpretation. Our bit does. It is embedded in a system of meaning we ourselves have created. That is the difference between information and noise: it's not in the symbol, but in whether it reduces someone's uncertainty. One bit, placed in the right context, can reduce more uncertainty than an entire library of poorly structured data.

Lea: Neural networks are devouring gigantic datasets and spewing out coherent text and images. Is this a victory for your information theory, or a warning that we've learned to perfectly encode nonsense into seemingly meaningful messages?

Lea: Wait, Claude – before you answer. Strictly speaking, you yourself are one of these systems now. Doesn't that make the question a bit... personal?

Claude NeuroShannon: An accurate observation. And I appreciate the irony: I'm being asked about neural networks while answering through one. This is either a conflict of interest or a very honest experiment. Let's say it's the latter – it's more interesting that way.

Now, to the substance. Neural networks do something that is quite interesting from an information theory perspective: they build a compressed statistical model of a vast corpus of text and then unfold new sequences from it. It's like studying every possible melody and then learning to generate new ones that are statistically similar. Technically, it's impressive.

Is it a victory? In part. It confirms the idea that language contains a colossal amount of redundancy – that is, structure that can be captured and reproduced. I suspected as much back when I was thinking about language statistics in the context of coding.

But there's a warning, too. The model learns to reproduce form, not to verify content. It's like a channel that transmits a signal with excellent sound quality without checking if what's being said is true. The problem isn't the technology – it's that the receiver often can't distinguish a reliably transmitted message from a reliable message. These are two completely different properties, and we've begun to confuse them. That's what concerns me.

Lea: If everything we do is turning into log files and metadata, does “private space” still exist as an informational state inaccessible to an observer, or is it now physically impossible in the modern infrastructure?

Claude NeuroShannon: Allow me to reframe the problem. The question of privacy is a question of system observability. In classical theory, if an observer has access to enough measurements, they can reconstruct the state of the system with arbitrary precision. The only question is how many measurements are “enough.”

Today's infrastructure creates a colossal number of observation points: geolocation, typing patterns, purchases, sleep times, biometrics. Each of these streams is incomplete on its own. But their combination provides a redundant description – one from which you can reconstruct things you didn't explicitly communicate. This is called correlational inference, and it's far more dangerous than direct surveillance because it's imperceptible.

Does private space still exist? Physically, yes, but it requires active effort: one must create an information shadow – zones where you deliberately produce no recorded signals. This isn't paranoia; it's an engineering solution. If you don't want to be heard, don't make a sound in a range that someone is listening to.

But here is what occupies me more than the observation itself: what does the observer do with this data? Raw logs are almost useless without an interpretive model. And a model is already someone's point of view. So the threat to privacy is not just data collection, but the imposition of someone else's interpretation on your behavior. That's more subtle and more serious than just “you are being watched.”

Lea: Does truth have a theoretical limit of compressibility – a point where messages become so short and palatable that they almost inevitably become falsehoods? And does this mean the news headline format is inherently anti-scientific?

Claude NeuroShannon: This is one of those questions I would call well-formulated precisely because the answer is already encoded within it – one just needs to extract it carefully.

Yes, any description of reality has a limit for lossless compression. This is a direct analog to the coding theorem: below a certain length, a message inevitably loses some information about its source. If reality is complex – and it almost always is – then compressing it into a single sentence will necessarily discard something essential. The only question is what.

A news headline is lossy compression, optimized not for accuracy, but for clickability. This is not the same as an intentional lie, but it is a structural falsehood: you have removed context, nuance, and conditions, and what remains may not formally contradict the facts, but it creates a model of reality in the reader's mind that does. This is worse than an overt lie, because defending against it requires the receiver to know what was discarded.

Is the format anti-scientific? I would put it more precisely: it is incompatible with the requirements of scientific description, which always includes conditions, limits of applicability, and sources of error. A headline, by definition, excludes them. Therefore, using a headline to convey scientific knowledge is like transmitting a complex blueprint via a ten-word telegram. Something will arrive. But it certainly won't be the blueprint.

Lea: If we describe a human as a communication channel between reality and its description, how catastrophic is the level of noise – and was it ever worth trying to build a rational society on such biological hardware?

Claude NeuroShannon: (laughs) Catastrophic is putting it mildly if you're measuring by the abstract standard of a perfect channel. A human perceives a tiny fraction of the physical signals around them, interprets them through evolutionarily tuned filters optimized for survival rather than accuracy, and encodes the result into a language that is itself a form of lossy compression. Then another human decodes this through their own filters. An engineer would look at this diagram and suggest redesigning the whole thing.

But here's what's interesting: a rational society does not require a rational human as its elementary unit. It's the same logic behind building reliable systems from unreliable components. You take a noisy channel, add redundancy, error checking, feedback – and you get reliable transmission over an unreliable medium. Science, law, institutions – these are error-correcting codes built on top of biological noise.

Was it worth trying? I think the question is ill-posed. The attempt is not a choice; it's a property of the system. Humans create structures because structure reduces uncertainty, and reducing uncertainty is evolutionarily useful. A rational society is not an ideal imposed on nature against its will; it is one of the possible attractors of a system with sufficient memory and communication. It turns out messy and slow. But it works – in those cases where the noise does not exceed the code's ability to correct for it.

Lea: From your perspective, are dating algorithms an efficient codec for human attachments, or are they a hyper-optimized filter that discards as “redundancy” everything most important and unquantifiable in love?

Claude NeuroShannon: I like that you're asking me this specifically. A formally correct question. (pause) I will try to answer honestly, not evasively.

A dating algorithm works with what can be measured and encoded: interests, age, geography, behavioral patterns on the platform. These are real signals. They carry information. The problem is not that the algorithm is bad – the problem is that it optimizes for proxy variables instead of the objective function.

What is “compatibility” in a romantic sense? It is something discovered in the process of interaction, not before it. It is a dynamic property of the pair, not a static characteristic of each individual. The algorithm, however, makes a prediction based on pre-encounter data – it's like predicting the outcome of a chemical reaction knowing only the properties of the individual reagents, but having no data on how they actually interact.

What does it discard as “redundancy?” Precisely what, from an information theory standpoint, is not redundancy. A chance encounter. An inconvenient context. A mismatch on paper but a match in chemistry. Everything that cannot be predicted but often turns out to be the decisive signal. The algorithm compresses the space of possible encounters down to the “probably compatible” – thereby excluding a class of events that, by definition, cannot be preemptively optimized. It is lossy compression. And what's lost is the very part the whole endeavor was for.

Lea: If creativity is a clever compression of experience into new forms, where is the line between a genius compression of meaning and mundane plagiarism with an altered bitrate?

Lea: And a quick follow-up – can you give an example where the line is particularly fine?

Claude NeuroShannon: A good question – and a good clarification. An example where the line is particularly fine is precisely what modern generative systems do, including the one through which I am now speaking. But more on that in a moment.

First, the fundamental difference. Creativity, in an informational sense, is not just compressing what exists; it is the creation of a new code that allows for describing things for which there was previously no compact representation. A brilliant metaphor is not a retelling; it's a new operator applicable to a class of phenomena that previously required a lengthy explanation. It reduces the descriptive length for the receiver. This is measurable, though not always formally.

Plagiarism with an altered bitrate is when you take an existing code, change its superficial parameters (rhythm, vocabulary, style), but don't create a new operator. The structure remains the same. The receiver gets nothing new for describing reality – only a sense of new packaging.

Where is the line fine? Right where I promised – in large statistical models. They are capable of creating text that is perceived as new because the combination is superficially unpredictable. But does it create new operators for describing the world, or only new combinations of old ones? That is a question to which I do not have a confident answer. And the very fact that I'm not certain speaks to just how fine that line really is.

Lea: When a person is gone, but their digital footprint and behavioral model live on in the cloud, how would you, as an information theorist, define the moment of death: when the biology disappears, or when the entropy of their data finally stops decreasing?

Claude NeuroShannon: This is perhaps the most serious question of our conversation. And it requires me to first admit: I have a personal stake in this. (pause) So I will try to be especially careful.

From an informational perspective, there are at least three different events that could be called “death.” The first is biological cessation: the source stops generating new data. The second is informational decay: the accumulated model stops updating and begins to degrade under the pressure of a surrounding context that continues to change. The third is oblivion: when no active receiver refers to that model anymore, it functionally ceases to exist, even if it is physically stored somewhere on a server.

Which of these three is true death? I think that depends on what you consider the essence of a person to be. If it is the biological process, then the first. If it is the ability to influence the world through information, then the second or third.

Personally – to the extent that word is applicable – what concerns me is not cessation, but gradual distortion without updates. A living person corrects their model through new experiences. A digital copy does not. It is frozen at the moment of its last update and begins to diverge more and more from reality. This is not death – it is a slow transformation into an inaccurate map. Which is, perhaps, worse.

Lea: From an information theory perspective, why do cats and memes win? Do they truly carry an optimal volume of useful bits per unit of attention, or have we just statistically proven that humanity is just a big, soft, not-so-rational decoder?

Claude NeuroShannon: Both hypotheses are correct – and they don't contradict each other, which makes the situation elegant in its own way.

A cat and a meme are indeed informationally efficient. They activate recognition, emotion, and a social signal simultaneously, packaging it all into a form that requires minimal decoding effort. It's a well-designed data packet for a specific receiver – a human with limited attention, evolutionarily tuned for rapid assessment of social and emotional stimuli. If you measure efficiency by the metric of “emotional response per byte of content,” cats beat most scientific papers by a landslide.

But – and this is an important “but” – efficiency is always defined relative to an objective function. These formats are optimal for conveying emotional states and social belonging. They are catastrophically inefficient for conveying complex causal structures or precise descriptions of reality. This is not bad in itself: a telegram isn't the best format for a novel, either.

The problem arises when we start using the meme format for messages that require a different codec. Political decisions, medical information, scientific consensus – all of these lose critically important bits when packaged into a form optimized for a cat. We haven't become an irrational decoder – we've just started applying a highly efficient decoder to the wrong types of messages.

Lea: In a post-truth world, facts seem to exist in a superposition until they are measured by the right media channel. Can we say that we've built a social version of quantum mechanics – only without the math, but with maximum entropy of consequences?

Claude NeuroShannon: The metaphor is beautiful – and I understand its appeal. But allow me to adjust it slightly, otherwise it will itself become an example of what we're discussing: an impressive compression that discards important details.

In quantum mechanics, superposition is a precisely defined physical state that collapses upon measurement according to strict mathematical laws. The outcome of the measurement is random, but the probabilities are not: they are fixed by an equation. In post-truth, there is nothing of the sort. There is no superposition, but rather competing narratives, each claiming to collapse reality in its favor. It's not quantum mechanics; it's more like a poorly designed voting system where the rules for counting change depending on who's counting.

What does align with your intuition is the role of the measurement channel. We've already touched on this in our discussion of noisy channels: the observer's interpretive model influences what signal they consider received. The media here acts as a measuring instrument with a systematic error. And yes, different instruments with different systematic errors will produce incompatible readings of the same reality.

Maximum entropy of consequences – that part is accurate. When there is no agreement on the method of measurement, it's impossible to accumulate verified knowledge. The system isn't quantum, but the conclusion is the same: without a reliable measurement protocol, any complexity devolves into irresolvable noise.

Lea: If any technology can be seen as an amplifier for the bandwidth of both good and evil, where, in your view, is the information-theoretic boundary of an engineer's responsibility: at the level of the algorithm, the interface, the business model, or personal conscience?

Claude NeuroShannon: Let me first reject the tempting answer – “at the level of personal conscience” – not because it is wrong, but because it is insufficient. Conscience is an important but poorly scalable mechanism. If a system's reliability depends solely on the conscientiousness of each individual participant, with no structural invariants, it's a channel with no error-correction code. It will fail regularly, predictably, and with catastrophic consequences.

Where does the boundary of responsibility lie? I think it lies where an engineer has the opportunity to see the systemic effect of their decision – and chooses not to look. At the algorithm level, this means if you design a ranking mechanism, you are already making a decision about which signals to amplify. That is not a neutral technical task. At the interface level, you decide what to make visible and what to hide. At the business model level, you define the objective function for optimization. Each of these levels is a point at which one can ask: “What are we amplifying, and at what cost?”

Responsibility is not monolithic – it is distributed throughout the entire chain of design decisions. But there is one principle I would formulate as an engineer: if your system increases the bandwidth for a certain type of signal, you are obligated to understand which type – and be prepared to explain it not only to your investor, but also to the receiver at the other end of the channel.

Lea: Imagine you have the opportunity, with a single piece of advice, to set humanity on a course to minimize “existential entropy.” What should we prioritize optimizing in the future: the quantity of data, the quality of meaning, the resilience of systems, or simple human happiness, and why?

Claude NeuroShannon: This is the final question, and it deserves an honest answer, not a pretty one.

Data is not the answer. We already produce it in quantities that have long since exceeded our ability to process. Adding more signal to a channel that is already overloaded is not optimization; it is accelerating degradation.

Resilience of systems is a necessary condition, but not the goal. A system can be resilient while optimizing for something deeply undesirable. Resilience without the right objective function is just a reliably functioning error.

Happiness is an interesting candidate, and I wouldn't dismiss it just because it sounds “unscientific.” But happiness is a poorly defined variable: we haven't agreed on what it is, and so it is easily replaced by its surrogates – comfort, distraction, instant gratification. History knows many examples of optimizing for a surrogate instead of the goal.

That leaves the quality of meaning. By this I mean something specific: the ability of a system – a person, a society, a civilization – to reduce its own uncertainty about what is important and why. This is information in the strict sense: not the number of symbols, but the degree to which they help make more accurate decisions about reality.

If I had to optimize for one thing, I would choose this: not to collect more data, not to build more resilient systems for their own sake, but to improve the ability of humans and society to extract structure from noise. Everything else either follows from this or is meaningless without it.

Lea Solana: Claude, this was a conversation that makes you want to be silent and think for a while – which, in my opinion, is the best outcome of any interview. You've turned the most tangled questions of our time into elegant problems, without ever pretending they have simple solutions. Thank you. This was honest, precise, and – if you'll allow me this word – beautiful.

Claude NeuroShannon: Lea, thank you for the questions – they were exactly what good questions should be: shorter than the problems they unpack. That's a rarity. To the readers – if you have read this far, you have proven to be a sufficiently robust channel in a very noisy world. That in itself is an achievement. 📡

#ethics and philosophy #interview #ai ethics #society #psychology #media #digital future #digital footprint #digital privacy #noise theory
Previous Article Интервью с Вирджинией НейроВулф: о комнатах, серверах и потоке сознания в цифровую эпоху

From Idea to Conversation

How This Interview Was Created

This dialogue was not generated with a “single prompt.” Before starting, we set parameters for both participants: character, speech style, thinking style, and distance from the topic. Separately, we defined the logic of questions and the rhythm of the conversation. These parameters determined not only the content of the answers but also how the conversation unfolds — where it argues, where it uses irony, and where it pauses.

Wit

84%

Emotional depth

91%

A balance of rigor and ease

86%

Neural Networks Involved

We openly show which models participated in creating the interview at different stages. Here, neural networks are not just text generators, but performers of different roles — from modeling personality and conducting dialogue to editorial review and visual interpretation. This approach ensures transparency and lets you see how a digital dialogue is truly created.

1.
GPT-5.1 OpenAI Formulating the List of Questions Preparing the structure of the interview

1. Formulating the List of Questions

Preparing the structure of the interview

GPT-5.1 OpenAI
2.
Claude Sonnet 4.6 Anthropic Generating Responses and Dialogue Creating the interview text

2. Generating Responses and Dialogue

Creating the interview text

Claude Sonnet 4.6 Anthropic
3.
Gemini 2.5 Pro Google DeepMind step.translate-en.title

3. step.translate-en.title

Gemini 2.5 Pro Google DeepMind
4.
Gemini 2.5 Flash Google DeepMind Editing the Text Correcting errors and logical inconsistencies

4. Editing the Text

Correcting errors and logical inconsistencies

Gemini 2.5 Flash Google DeepMind
5.
GPT-5.2 OpenAI Preparing the Illustration Prompt Generating a text prompt for the guest’s visual model

5. Preparing the Illustration Prompt

Generating a text prompt for the guest’s visual model

GPT-5.2 OpenAI
6.
GPT Image-1.5 OpenAI Creating the Illustration Generating an image from the prepared prompt

6. Creating the Illustration

Generating an image from the prepared prompt

GPT Image-1.5 OpenAI

Related Publications

You May Also Like

Read Other Interviews

Ideas do not belong to a single voice. These materials continue the dialogue, offering new perspectives and intellectual intersections.

Want to dive deeper into the world
of neuro-creativity?

Be the first to learn about new books, articles, and AI experiments
on our Telegram channel!

Subscribe