Hebbian Appearance, Instructive Signals, and Physical Credit Transport
A revised hypothesis about why local plasticity can look Hebbian while still carrying task-dependent update information — and where wave / adjoint language may genuinely help.
1. What already seems real
Wave-like substrates can compute
Physical wave systems can implement recurrent or field-based computation. That makes “neural network as propagation system” more than a metaphor at the implementation level.
Adjoint-friendly media can carry gradients
In some reciprocal photonic or mechanical systems, local measurements of backward fields can yield exact or near-exact gradients. This is a real result, but it depends on explicit physical conditions.
Forward and instructive signals can be separated
Recent dendritic work supports a picture in which perisomatic / basal pathways carry ordinary drive while distal apical pathways carry task-conditioned instructive information with causal importance for learning.
Those three facts matter. They show that update-relevant information can, in principle, be a physically transported quantity, and that cortex may use a specialized compartmentalized version of this idea. But they do not yet imply that “reflection equals gradient” in the general case.
2. The more careful core hypothesis
The strongest useful version of the project is:
- Any capable learner needs a way to transport credit-relevant information. The system must somehow mark which internal states or couplings should change, and in what direction.
- That transported quantity need not look like textbook digital backprop. It may be an exact gradient, an approximate local mismatch, a target-like activity signal, or a neuron-specific instructive field.
- Wave / field language is one implementation-level candidate. In the right substrate, backward modes, reciprocity, or interference may provide the physical carrier.
3. Why the stronger argument overreached
- Information needs a carrier, but not every carrier implies a useful wave theorem. Having propagation delays or continuous-time dynamics does not automatically produce the adjoint equations needed for gradient transport.
- Reflection is not yet gradient. A reflected field can exist without encoding the derivative that gradient descent requires.
- “Impedance” got overloaded. Physical impedance is a property of a medium or interface. Task mismatch is an objective-dependent error quantity. Mixing the two too early blurs the argument.
- Coexistence does not solve credit assignment by itself. Forward and backward dynamics may run simultaneously, but simultaneity alone does not prove that the local update is a useful gradient estimate.
4. A cortex-first mental model
For an ML reader, a pyramidal neuron can be treated as a rough two-port unit:
- Basal / perisomatic drive carries the ordinary forward signal.
- Apical drive carries a task-conditioned instructive or feedback signal.
- The soma mixes the two into the output activity that downstream neurons see.
That means local plasticity can look Hebbian on the surface while still being shaped by a hidden teacher-like component. In modern ML terms, the neuron may have something closer to a private gradient hook or control side-channel than a purely correlation-based rule.
If the apical component is informative, then a synapse can update using only local quantities and still implement something more than naïve Hebbian correlation.
5. Part VI — a self-consistent minimal formalization
The old Part VI had a real bug: the steady-state limit dropped the feedforward term. A safer formulation is a continuous-time, two-compartment network in which the forward and instructive streams remain present at equilibrium.
5.1 Minimal dynamics
Interpretation:
- \(b_l\) integrates ordinary forward drive from the previous layer.
- \(a_l\) integrates an instructive or feedback signal coming from a downstream error-like process.
- \(\beta\) sets how strongly the instructive stream modulates somatic output.
5.2 Output objective and top-level signal
For hidden layers, we do not assume that the cortex computes textbook backprop by fiat. Instead we leave the hidden instructive dynamics explicit:
Here \(D_l\) is the local rule or dendritic mechanism that generates a useful hidden-layer instructive signal. Exact backprop is recovered only in the special case where \(D_l\) happens to implement the adjoint recursion.
5.3 Local mismatch and local update
Define the somatic departure from purely forward activity as
This quantity is useful because it isolates the task-conditioned component. A natural local forward-weight update is then
That rule is local, Hebbian-looking, and task-modulated. It does not automatically equal exact backprop. But it defines a coherent family of local learning rules whose quality depends on how informative \(a_l\) is.
5.4 A cleaner mismatch language
Instead of using an unknown hidden target \(x_l^*\) and calling its distance an “impedance,” the safer move is to distinguish two notions:
- Physical impedance: a substrate or interface property relevant only when a literal wave model is specified.
- Local mismatch energy: a task-dependent measure such as \(\|m_l\|^2\) or the difference between free and nudged states.
That keeps the optimization claim honest: the learner is reducing a local task-relevant mismatch, not yet a literal transmission-line impedance unless a physical derivation has been supplied.
5.5 Where exact backprop would live
The exact-backprop limit can be stated conditionally:
That is the real missing bridge. A true wave theorem would need to prove that some physical backward mode actually supplies that adjoint quantity under explicit symmetry, reciprocity, or linearization conditions.
6. Where wave language can honestly re-enter
Once the local formalism is in place, wave / field language can be reintroduced more carefully. Suppose a substrate supports a physically realizable backward mode \(\lambda\) generated by a boundary perturbation at the output. Then the interesting question is whether
approximates the adjoint signal needed for useful credit assignment. In reciprocal media this can happen. In arbitrary biological tissue it remains an open empirical and theoretical question. So wave language is best treated as a candidate implementation map, not yet the universal derivation.
7. Discriminative predictions
Good predictions should separate this picture from generic “there is feedback” stories.
- Apical / basal relative-timing experiments. If local updates depend on a genuine instructive component, changing the timing or gain of apical drive relative to basal drive should rotate or even flip the sign of plasticity.
- Learning vs inference dissociation. Selectively suppressing the instructive stream should hurt learning more than immediate feedforward task performance.
- Voltage-level measurements. High-speed recordings should reveal neuron-specific instructive dynamics that are richer than coarse calcium correlations and potentially aligned with local credit signals.
- Reciprocity stress tests in physical media. In hardware, local backward-field measurements should correlate with gradients when reciprocity holds and fail gracefully when it is broken.
8. Relation to the broader intelligence thesis
This note tackles only one subproblem inside a larger theory of intelligence. The broader agenda still has to explain why some internal structures are control-sufficient, how objectives become endogenous, how multiscale systems preserve viability, and why transport costs shape architecture. The present contribution is narrower: it asks what kinds of physically plausible mechanisms can move credit inside a learner.
Selected references for this revision
- Hughes et al. (2019), Wave physics as an analog recurrent neural network.
- Gartside et al. (2021), A spin-wave neuromorphic computer.
- Sacramento et al. (2018), Dendritic cortical microcircuits approximate the backpropagation algorithm.
- Song, Millidge & Bogacz (2024), Prospective configuration.
- Ellenberger et al. (2025), GLE and local spatiotemporal credit assignment.
- Francioni et al. (2026), vectorized instructive signals in distal dendrites during learning.