| ▲ | thecr0w a day ago | |||||||
Hm, I didn't try exactly this, but I probably should! Wrt unit test script, let's take Claude out of the equation, how would you design the unit test? I kept running into either Claude or some library not being capable of consistently identifying planet vs non planet which was hindering Claude's ability to make decisions based on fine detail or "pixel coordinates" if that makes sense. | ||||||||
| ▲ | yfontana 9 hours ago | parent | next [-] | |||||||
If I were to do this (and I might give it a try, this is quite an interesting case), I would try to run a detection model on the image, to find bounding boxes for the planets and their associated text. Even a small model running on CPU should be able to do this relatively quickly. | ||||||||
| ▲ | cfbradford a day ago | parent | prev [-] | |||||||
Do you give Claude the screenshot as a file? If so I’d just ask it to write a tool to diff each asset to every possible location in the source image to find the most likely position of each asset. You don’t really need recognition if you can brute force the search. As a human this is roughly what I would do if you told me I needed to recreate something like that with pixel perfect precision. | ||||||||
| ||||||||