Remix.run Logo
pncnmnp 3 days ago

I have a question that's bothered me for quite a while now. In 2018, Michael Jordan (UC Berkeley) wrote a rather interesting essay - https://medium.com/@mijordan3/artificial-intelligence-the-re... (Artificial Intelligence — The Revolution Hasn’t Happened Yet)

In it, he stated the following:

> Indeed, the famous “backpropagation” algorithm that was rediscovered by David Rumelhart in the early 1980s, and which is now viewed as being at the core of the so-called “AI revolution,” first arose in the field of control theory in the 1950s and 1960s. One of its early applications was to optimize the thrusts of the Apollo spaceships as they headed towards the moon.

I was wondering whether anyone could point me to the paper or piece of work he was referring to. There are many citations in Schmidhuber’s piece, and in my previous attempts I've gotten lost in papers.

drsopp 3 days ago | parent | next [-]

Perhaps this:

Henry J. Kelley (1960). Gradient Theory of Optimal Flight Paths.

[1] https://claude.ai/public/artifacts/8e1dfe2b-69b0-4f2c-88f5-0...

pncnmnp 3 days ago | parent | next [-]

Thanks! This might be it. I looked up Henry J. Kelley on Wikipedia, and in the notes I found a citation to this paper from Stuart Dreyfus (Berkeley): "Artificial Neural Networks, Back Propagation and the Kelley-Bryson Gradient Procedure" (https://gwern.net/doc/ai/nn/1990-dreyfus.pdf).

I am still going through it, but the latter is quite interesting!

cco 3 days ago | parent | prev [-]

Count another in the win column for the USA's heavy investment into basic sciences during the space race.

So sad to see the current state. Hopefully we can turn it around.

leokoz8 3 days ago | parent | prev | next [-]

It is in Applied Optimal Control by Bryson and Ho (1969). Yann LeCun acknowledges this in his 1989 paper on backpropagation:https://new.math.uiuc.edu/MathMLseminar/seminarPapers/LeCunB....

> "Since his first work on the subject, the author has found that A. Bryson and Y.-C. Ho [Bryson and Ho, 1969] described the backpropagation algorithm using Lagrange formalism. Although their description was, of course, within the framework of optimal control rather than machine learning, the resulting procedure is identical to backpropagation."

mellosouls 3 days ago | parent | prev | next [-]

See Widnall's overview here which discusses some of the ground that crosses over with what has come to be known as backpropagation:

The Minimum-Time Thrust-Vector Control Law in the Apollo Lunar-Module Autopilot (1970)

https://www.sciencedirect.com/science/article/pii/S147466701...

psYchotic 3 days ago | parent | prev | next [-]

I found this,maybe it helps: https://gwern.net/doc/ai/nn/1986-rumelhart-2.pdf

pncnmnp 3 days ago | parent [-]

Apologies - I should have been clear. I was not referring to Rumelhart et al., but to pieces of work that point to "optimizing the thrusts of the Apollo spaceships" using backprop.

observationist 3 days ago | parent | next [-]

Kelley 1960 (the gradient/adjoint flight‑path paper) https://perceptrondemo.com

AIAA 65‑701 (1965) “optimum thrust programming” for lunar transfers via steepest descent (Apollo‑era) https://arc.aiaa.org/doi/abs/10.2514/6.1965-701

Meditch 1964 (optimal thrust programming for lunar landing) https://openmdao.github.io/dymos/examples/moon_landing/moon_...

Smith 1967 & Colunga 1970 (explicit Apollo‑type trajectory/re‑entry optimization using adjoint gradients) https://ntrs.nasa.gov/citations/19670015714

One thing AI has been great for, recently, has been search for obscure or indirect references like this, that might be one step removed from any specific thing you're searching for, or if you have a tip-of-the-tongue search where you might have forgotten a phrase, or know you're using the wrong wording.

It's cool that you can trace the work of these rocket scientists all the way to the state of the art AI.

costates-maybe 3 days ago | parent | prev [-]

I don't know if there is a particular paper exactly, but Ben Recht has a discussion of the relationship between techniques in optimal control that became prominent in the 60's, and backpropagation:

https://archives.argmin.net/2016/05/18/mates-of-costate/

seertaak 3 days ago | parent | prev | next [-]

Rumelhart et al wrote "Parallel Distributed Processing"; there's a chapter where he proves that the backprop algorithm maximizes "harmony", which is simply a different formulation of error minimization.

I remember reading this book enthusiastically back in the mid 90s. I don't recall struggling with the proof, it was fairly straightforward. (I was in senior high school year at the time.)

duped 3 days ago | parent | prev | next [-]

They're probably talking about Kalman Filters (1961) and LMS filters (1960).

pjbk 3 days ago | parent [-]

To be fair, any multivariable regulator or filter (estimator) that has a quadratic component (LQR/LQE) will naturally yield a solution similar to backpropagation when an iterative algorithm is used to optimize its cost or error function through a differentiable tangent space.

bgnn 3 days ago | parent [-]

So yeah, this was what I was thinking for a while. What about a more nonlinear estimator? Intuitively seems similar to me.

andyferris 3 days ago | parent [-]

I believe the reason it works in nonlinear cases is that the derivative is “naturally linear” (to calculate the derivative, you are considering ever smaller regions where the cost function is approximately linear - exactly “how nonlinear” the cost function is elsewhere doesn’t play a role).

bgnn 3 days ago | parent [-]

that makes a lot of sense actually. thank you.

aaron695 3 days ago | parent | prev | next [-]

[dead]

dataflow 3 days ago | parent | prev | next [-]

[flagged]

throawayonthe 3 days ago | parent [-]

it's rude to show people your llm output

dataflow 3 days ago | parent | next [-]

Wow, this is the first time I'm hearing such a thing. For clarity:

I pasted the output so a ton of people wouldn't repeat the same question to ChatGPT and burn a ton of CO2 to get the same answer.

I didn't paste the query since I didn't find it interesting.

And I didn't fact check because I didn't have the time. I was walking and had a few seconds to just do this on my phone.

Not sure how this was rude, I certainly didn't intend it to be...

epaulson 3 days ago | parent [-]

The 'it's rude to show your ai(llm) output' is a reference to this: https://distantprovince.by/posts/its-rude-to-show-ai-output-...

dataflow 3 days ago | parent [-]

What is "this" exactly? Is it a well-known author or website? Or otherwise a reference that one should be familiar with? It looks like a random blog to me... with an opinion declared as fact that's quite easy to refute.

cubefox 3 days ago | parent | next [-]

It got 321 points here: https://news.ycombinator.com/item?id=44617172

actionfromafar 3 days ago | parent | prev [-]

”This” is also the novel Blindsight.

drsopp 3 days ago | parent | prev | next [-]

Why?

danieldk 3 days ago | parent [-]

Because it is terribly low-effort. People are here for interesting and insightful discussions with other humans. If they were interested in unverified LLM output… they would ask an LLM?

drsopp 3 days ago | parent [-]

Who cares if it is low effort? I got lots of upvotes for my link to Claude about this, and pncnmnp seems happy. The downvoted comment from ChatGPT was maybe a bit spammy?

lcnPylGDnU4H9OF 3 days ago | parent | next [-]

> Who cares if it is low effort?

It's a weird thing to wonder after so many people expressed their dislike of the upthread low-effort comment with a down vote (and then another voiced a more explicit opinion). The point is that a reader may want to know that the text they're reading is something a human took the time to write themselves. That fact is what makes it valuable.

> pncnmnp seems happy

They just haven't commented. There is no reason to attribute this specific motive to that fact.

drsopp 3 days ago | parent [-]

> The point is that a reader may want to know that the text they're reading is something a human took the time to write themselves.

The reader may also simply want information that helps them.

> They just haven't commented.

Yes, they did.

Dylan16807 3 days ago | parent [-]

> The reader may also simply want information that helps them.

The reader will generally want at least a cursory verification that it is information that helps, which dataflow didn't try to do.

Especially when you're looking for specific documents and you don't check if the documents are real. (dataflow's third one doesn't appear to be.)

drsopp 3 days ago | parent [-]

This I agree with completely.

bee_rider 3 days ago | parent | prev [-]

Yours was a little bit more useful, it you essentially used the LLM as a search engine to find a real article, right?

Directly posting the random text generated by the LLM is more annoying. I mean, they didn’t even vouch or it or verify that it was right.

aeonik 3 days ago | parent | prev [-]

I don't think it's rude, it saves me from having to come up with my own prompt and wade through the back and forth to get useful insight from the LLMs, also saves me from spending my tokens.

Also, I quite love it when people clearly demarcate which part of their content came from an LLM, and specifies which model.

The little citation carries a huge amount of useful information.

The folks who don't like AI should like it too, as they can easily filter the content.

cubefox 3 days ago | parent | prev [-]

> ... first arose in the field of control theory in the 1950s and 1960s. One of its early applications was to optimize the thrusts of the Apollo spaceships as they headed towards the moon.

I think "its" refers to control theory, not backpropagation.