Remix.run Logo
jarjoura 4 hours ago

You do realize though that using Claude Desktop to "search" through confluence is like paying a world class architect on the hour just to give you some tips on how to layout your small loft to maximize sunlight.

This is such a perfect example of the mania behind this rollout.

There's no way you can make the financials work here compared to JetBrains spending the same millions spent on AI infrastructure and instead building better search in Confluence. Confluence search SUCKS, but that's just a lack of focus (or resources) on building a more complex, more robust solution. It's a wiki.

Either way, making a more robust search is a one time cost that benefits everyone. Instead, you're running a piece of software that goes directly to Anthropic's bank account, and to the data centers and to hyper scalers. Every single query must be re-run from scratch, costing your company a fortune, that if not managed properly will come out of spending that money elsewhere.

xboxnolifes 4 hours ago | parent | next [-]

> You do realize though that using Claude Desktop to "search" through confluence is like paying a world class architect on the hour just to give you some tips on how to layout your small loft to maximize sunlight.

If I could pay a world class architect $1.50 to give me tips on how to maximize sunlight in my loft I would.

Would it be nice if confluence just had a robust search that had a one time cost and then benefited everyone thereafter? Sure, but that's not the current reality, and I do not have control over their actions. I can only control mine.

lukevp 4 hours ago | parent | prev [-]

And what is using Confluence in the first place? Your MacBook Pro is faster than a supercomputer from 20 years ago. As we make compute cheaper, we find ways to use it that are less efficient in an absolute sense but more efficient for the end user. A graphical docs portal like Confluence is a hell of a lot easier to use than EMacs and SSH to edit plain text files on an 80 character terminal. But it uses thousands of times more compute.

It seems ridiculous right now because we don’t have hardware to accelerate the LLMs, but in 5 years this will be trivial to run.

jarjoura an hour ago | parent [-]

I'm confused by your analogy. A wiki run server is extremely efficient to run, and can be hosted from a tiny little raspberry pie. A search engine can be optimized to provide results near O(1). You can even pull up and read results on a very old computer. All of the concerns around cost and resource efficiency can be addressed as all of this is a solved problem.

Even with an LLM agent getting cheaper to run in the future, it's still fundamentally non-deterministic so the ongoing cost for a single exploration query run can never get anywhere near as cheap as running a wiki with a proper search engine.