<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Telecommunications on Corebaseit — POS · EMV · Payments · AI</title><link>https://corebaseit.com/tags/telecommunications/</link><description>Recent content in Telecommunications on Corebaseit — POS · EMV · Payments · AI</description><generator>Hugo -- gohugo.io</generator><language>en-us</language><managingEditor>contact@corebaseit.com (Vincent Bevia)</managingEditor><webMaster>contact@corebaseit.com (Vincent Bevia)</webMaster><lastBuildDate>Tue, 14 Apr 2026 10:00:00 +0100</lastBuildDate><atom:link href="https://corebaseit.com/tags/telecommunications/index.xml" rel="self" type="application/rss+xml"/><item><title>I Spent Years on Adaptive Filters. I Was Already Training Neural Networks.</title><link>https://corebaseit.com/corebaseit_posts_in_review/lms-adaptive-filters-and-neural-network-training/</link><pubDate>Tue, 14 Apr 2026 10:00:00 +0100</pubDate><author>contact@corebaseit.com (Vincent Bevia)</author><guid>https://corebaseit.com/corebaseit_posts_in_review/lms-adaptive-filters-and-neural-network-training/</guid><description>&lt;p>&lt;strong>I spent years implementing LMS-based equalizers and echo cancellers in telecommunications. Only later did I fully appreciate what I had been doing mathematically: the same family of update rules that powers neural network training today.&lt;/strong>&lt;/p>
&lt;p>Not as a loose analogy — as the same structure of optimization. Widrow and Hoff formalized the Least Mean Squares (LMS) algorithm in 1960 for the Adaline. Rumelhart, Hinton, and Williams scaled related ideas through multi-layer networks with backpropagation in 1986. The vocabulary changed from &lt;em>adaptive filtering&lt;/em> to &lt;em>deep learning&lt;/em>, but the core idea — adjust parameters in the direction that reduces error, one small step at a time — is continuous across both worlds.&lt;/p>
&lt;p>This post is my attempt to make that lineage explicit: what LMS actually is, why it is structurally the same rule as stochastic gradient descent on a linear model, how the engineering trade-offs line up, and why non-stationarity remains the hard problem in both domains.&lt;/p>
&lt;hr>
&lt;h2 id="lms-is-not-a-metaphor-for-training--it-is-the-algorithm">LMS Is Not a Metaphor for Training — It Is the Algorithm
&lt;/h2>&lt;p>The LMS update for a linear combiner (FIR filter or single Adaline) is:&lt;/p>
&lt;p>$$
\mathbf{w}(n+1) = \mathbf{w}(n) + \mu , e(n) , \mathbf{x}(n)
$$&lt;/p>
&lt;p style="text-align: center;">
&lt;img src="https://corebaseit.com/diagrams/LMS_SGD_structural_equivalence_diagram.png" alt="LMS = SGD structural equivalence diagram" style="max-width: 900px; width: 100%;" />
&lt;/p>
&lt;p>Here (\mathbf{w}(n)) is the weight vector at time (n), (\mathbf{x}(n)) is the input vector (tap-delay line or feature vector), (e(n) = d(n) - y(n)) is the error between the desired response (d(n)) and the output (y(n) = \mathbf{w}^\top(n)\mathbf{x}(n)), and (\mu) is the step size.&lt;/p>
&lt;p>That is &lt;strong>stochastic gradient descent&lt;/strong> on the instantaneous squared error (\frac{1}{2}e^2(n)) with respect to (\mathbf{w}). The gradient of (\frac{1}{2}(d - \mathbf{w}^\top\mathbf{x})^2) with respect to (\mathbf{w}) is (-e,\mathbf{x}). Walking in the opposite direction of the gradient (or equivalently, in the direction (+e,\mathbf{x}) when you define the update as above) is exactly the LMS rule.&lt;/p>
&lt;p>So if you have ever shipped an LMS equalizer or echo canceller, you have implemented the foundational learning rule that underlies a huge fraction of modern machine learning: &lt;strong>small steps proportional to error times input&lt;/strong>. The notation in Haykin&amp;rsquo;s &lt;em>Adaptive Filter Theory&lt;/em> differs from PyTorch docs; the mathematics does not.&lt;/p>
&lt;p>Multi-layer networks add the chain rule (backpropagation) to compute how error propagates to earlier layers, but the &lt;strong>local&lt;/strong> update at a linear layer trained with mean squared error is still the same structural move: adjust weights in proportion to error and activations. Everything else — momentum, Adam, adaptive learning rates — is engineering on top of that spine.&lt;/p>
&lt;hr>
&lt;h2 id="the-engineering-trade-offs-are-the-same-trade-offs">The Engineering Trade-Offs Are the Same Trade-Offs
&lt;/h2>&lt;p>In telecommunications, the step size (\mu) controls the classic compromise: &lt;strong>convergence speed versus steady-state misadjustment&lt;/strong>. Too large — the filter can diverge or oscillate. Too small — the filter cannot track a fast-fading channel or a moving echo path. Entire chapters of adaptive filtering textbooks are devoted to stability bounds on (\mu) (often expressed in terms of input power and filter length) and to variants that fix the worst-case behavior.&lt;/p>
&lt;p style="text-align: center;">
&lt;img src="https://corebaseit.com/diagrams/Step_size_learning_rate_trade-off_diagram.png" alt="Step size / learning rate trade-off diagram" style="max-width: 900px; width: 100%;" />
&lt;/p>
&lt;p>In deep learning, the learning rate (\eta) plays the same role at a higher level: too high and training diverges or chatters around a minimum; too low and you underfit or burn compute without making progress. The community talks about learning-rate schedules, warm-up, and cosine decay — different names for the same instinct: &lt;strong>the right step size depends on the landscape and may need to change over time&lt;/strong>.&lt;/p>
&lt;p>&lt;strong>Normalized LMS (NLMS)&lt;/strong> scales the update by the inverse of the input energy (|\mathbf{x}(n)|^2) (with a small regularizer to avoid division by zero). The goal is stable convergence when input power varies — the same motivation that shows up in adaptive optimizers that normalize updates by running statistics of gradients (RMSProp-style normalization is not identical to NLMS, but the &lt;em>intent&lt;/em> — tame the step when the signal scale changes — is shared). The DSP community spent decades refining these ideas for real-time hardware; ML rediscovered many of the same pressures when training became unstable at scale.&lt;/p>
&lt;hr>
&lt;h2 id="non-stationarity-was-always-the-real-problem--and-still-is">Non-Stationarity Was Always the Real Problem — and Still Is
&lt;/h2>&lt;p>Adaptive filters were built for &lt;strong>non-stationary&lt;/strong> environments: multipath fading, time-varying echoes, drifting noise floors. The “true” optimal weights are not fixed; they move. The filter is not supposed to converge once and freeze — it is supposed to &lt;strong>track&lt;/strong>. That mindset is closer to production ML than a static batch fit on a fixed dataset.&lt;/p>
&lt;p>Modern systems face the same phenomenon under different labels: &lt;strong>distribution shift&lt;/strong>, &lt;strong>concept drift&lt;/strong>, stale features, changing user behavior, adversarial drift in inputs. The model that was optimal last month is not guaranteed to be optimal this month. Retraining on a schedule, online updates, monitoring, and guardrails are the engineering response — conceptually in the same family as “never assume the channel is static.”&lt;/p>
&lt;p>Research on in-context learning in linear models (for example Akyürek et al., 2022) even investigates which learning algorithms are implicitly approximated by transformers under simplified settings — another reminder that the boundary between classical adaptive signal processing and contemporary ML is thinner than course catalogs suggest.&lt;/p>
&lt;hr>
&lt;h2 id="the-bigger-picture">The Bigger Picture
&lt;/h2>&lt;p style="text-align: center;">
&lt;img src="https://corebaseit.com/diagrams/Historical_lineage_timeline_diagram.png" alt="Historical lineage / timeline diagram" style="max-width: 900px; width: 100%;" />
&lt;/p>
&lt;p>For engineers who came up through &lt;strong>telecommunications and signal processing&lt;/strong>, the move into AI is often described as a career pivot. In my experience, it is closer to a &lt;strong>change of vocabulary&lt;/strong> on top of a continuous mathematical thread: error-driven updates, step-size discipline, stability under non-stationarity, and the centrality of second-order statistics (explicitly in LMS, implicitly in much of modern training).&lt;/p>
&lt;p>The boundary between DSP and machine learning was never as sharp as the literature implied. If you understand LMS, you already understand a piece of what every deep learning framework is doing when it steps the weights. The rest is scale, architecture, and tooling — important, but not magic.&lt;/p>
&lt;hr>
&lt;h2 id="references">References
&lt;/h2>&lt;ul>
&lt;li>Widrow, B., &amp;amp; Hoff, M. E. &amp;ldquo;Adaptive switching circuits.&amp;rdquo; &lt;em>IRE WESCON Convention Record&lt;/em>, 4, 96–104, 1960.&lt;/li>
&lt;li>Haykin, S. &lt;em>Adaptive Filter Theory&lt;/em> (4th ed.). Prentice Hall, 2002.&lt;/li>
&lt;li>Rumelhart, D. E., Hinton, G. E., &amp;amp; Williams, R. J. &amp;ldquo;Learning representations by back-propagating errors.&amp;rdquo; &lt;em>Nature&lt;/em>, 323, 533–536, 1986.&lt;/li>
&lt;li>Akyürek, E. et al. &amp;ldquo;What learning algorithm is in-context learning? Investigations with linear models.&amp;rdquo; 2022. &lt;a class="link" href="https://arxiv.org/abs/2211.15661" target="_blank" rel="noopener"
>arxiv.org/abs/2211.15661&lt;/a>&lt;/li>
&lt;/ul>
&lt;h2 id="further-reading">Further reading
&lt;/h2>&lt;ul>
&lt;li>&lt;a class="link" href="https://corebaseit.com/posts/stochastic-entropy-ai/" >Stochastic, Entropy &amp;amp; AI: From Thermodynamics to Information Theory to Modern Machine Learning&lt;/a> — related thread on probability, information, and ML foundations&lt;/li>
&lt;li>&lt;em>The Obsolescence Paradox: Why the Best Engineers Will Thrive in the AI Era&lt;/em> — engineering judgment as tools and vocabulary change&lt;/li>
&lt;/ul></description></item></channel></rss>