Remix.run Logo
bfeynman 4 hours ago

I feel bad that people have to read this. It's complete puffery, made up for clicks, and the biggest thing is the pure bravado with which a company says, "Hey, let's just waste a ton of money, all for a potential blog and marketing piece." This is not really automated in any fashion. I was dubious at first, but then I saw the screencaps showing the devs interacting with Luna via a Slack workflow with a human in the loop — meaning they're literally just proxying their own behavior through an LLM. This is no different than anyone who consults AI for any decision with context. To get even more technical on the fallacy: this is not automation, as there is data leakage at every step where there is a human in the loop. A broken clock is right twice a day; an LLM could cycle through 100 guesses to pick a number, but don't market that as an oracle. Aside from that, you could just look at the pictures and context (retail in SF) and assume making a profit here would be near impossible. An actual AI ceo would probably have immediately cancel the lease.

insane_dreamer 2 hours ago | parent | next [-]

> I was dubious at first, but then I saw the screencaps showing the devs interacting with Luna via a Slack workflow with a human in the loop — meaning they're literally just proxying their own behavior through an LLM. This is no different than anyone who consults AI for any decision with context.

A human can be in the loop if the human is exactly executing the orders of the AI. It's still the AI making all the decisions, which is the purpose of the experiment - not to see whether agents can handle every interaction necessary to run a business (pick up the phone and place orders, etc.). That's also why Luna hired humans.

bfeynman 2 hours ago | parent [-]

that is ... not correct? This is classic example of data leakage, the yes/no things are signals feeding back to the model influencing (and here, basically guiding) future decisions.

insane_dreamer an hour ago | parent [-]

It's not data leakage.

If the experiment is to see how the AI behaves on its own, then of course it needs to know the outcomes of its decisions (either automatically, or fed to it by a human), which of course influence its next decisions. This is providing the AI with retained memory, which is essential to the experiment. It's similar to an AI writing code which it then runs and parses the logs to see the outcome and make improvements to it. (It is not _retrained_ on those outcomes, and neither is that the case here; but it can reference them in stored memory.)

j2kun 3 hours ago | parent | prev | next [-]

[flagged]

antonvs 2 hours ago | parent | next [-]

I appreciated the analysis given by the other commenter, so I'm glad they didn't take that lazy way out.

kryogen1c 3 hours ago | parent | prev | next [-]

The submitter appears to be a co-founder of the company the article is about (omitted from the HN account bio), and the article is misleading to the point of lying.

This company now has strong a strong negative reputation in my mind that I will gladly share with others.

themafia 3 hours ago | parent | prev [-]

This is Hacker News. It should be filled with curious people who are willing to express their opinions and points of view. To tell someone to just punitively flag something and then "move on" is absurdly reductive and small minded.

graybeardhacker 3 hours ago | parent | prev [-]

A stopped clock is right twice a day; a broken one can be wrong forever. Just saying.