| ▲ | PaulHoule 6 hours ago | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
My first take is that you could have 10 TB of logs with just a few unique lines that are actually interesting. So I am not thinking "Wow, what impressive big data you have there" but rather "if you have an accuracy of 1-10^-6 you are still are overwhelmed with false positives" or "I hope your daddy is paying for your tokens" | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| ▲ | aluzzardi 6 hours ago | parent | next [-] | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Mendral co-founder and post author here. I agree with your statement and explained in a few other comments how we're doing this. tldr: - Something happens that needs investigating - Main (Opus) agent makes focused plan and spawns sub agents (Haiku) - They use ClickHouse queries to grab only relevant pieces of logs and return summaries/patterns This is what you would do manually: you're not going to read through 10 TB of logs when something happens; you make a plan, open a few tabs and start doing narrow, focused searches. | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| ▲ | jcgrillo 6 hours ago | parent | prev [-] | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Yeah this is my experience with logs data. You only actually care about O(10) lines per query, usually related by some correlation ID. Or, instead of searching you're summarizing by counting things. In that case, actually counting is important ;). In this piece though--and maybe I need to read it again--I was under the impression that the LLM's "interface" to the logs data is queries against clickhouse. So long as the queries return sensibly limited results, and it doesn't go wild with the queries, that could address both concerns? | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||