▲ | astrange 2 days ago | |||||||||||||||||||||||||
> The easiest way of solving math problems with an LLM is to make sure that very similar problems are included in the training set. An irony here is that math blogs like Tao's might not be in LLM training data, for the same reason they aren't accessible to screen readers - they're full of math, and the math is rendered as images, so it's nonsense if you can't read the images. (The images on his blog do have alt text, but it's just the LaTeX code, which isn't much better.) | ||||||||||||||||||||||||||
▲ | alansammarone 2 days ago | parent | next [-] | |||||||||||||||||||||||||
As others have pointed out, LLMs have no trouble with LaTeX. I can see why one might think they're not - in fact, I made the same assumption myself sometime ago. LLMs, via transformers, are exceptionally good any _any_ sequence or one-dimensional data. One very interesting (to me anyway) example is base64 - pick some not-huge sentence (say, 10 words), base64-encode it, and just paste it in any LLM you want, and it will be able to understand it. Same works with hex, ascii representation, or binary. Here's a sample if you want to try: aWYgYWxsIEEncyBhcmUgQidzLCBidXQgb25seSBzb21lIEIncyBhcmUgQydzLCBhcmUgYWxsIEEncyBDJ3M/IEFuc3dlciBpbiBiYXNlNjQu I remember running this experiment some time ago in a context where I was certain there was no possibility of tool use to encode/decode. Nowadays, it can be hard to certain whether there is any tool use or not, in some cases, such as Mistral, the response is quick enough to make it unlikely there's any tool use. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
▲ | prein 2 days ago | parent | prev | next [-] | |||||||||||||||||||||||||
What would be a better alternative than LaTex for the alt text? I can't think of a solution that makes more sense, it provides an unambiguous representation of what's depicted. I wouldn't think an LLM would have issue with that at all. I can see how a screen reader might, but it seems like the same problem faced by a screen reader with any piece of code, not just LaTex. | ||||||||||||||||||||||||||
▲ | mbowcut2 a day ago | parent | prev | next [-] | |||||||||||||||||||||||||
LLMs are better at LaTeX than humans. ChatGPT often writes LaTeX responses. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
▲ | QuesnayJr 2 days ago | parent | prev | next [-] | |||||||||||||||||||||||||
LLMs understand LaTeX extraordinarily well. | ||||||||||||||||||||||||||
▲ | constantcrying a day ago | parent | prev | next [-] | |||||||||||||||||||||||||
>(The images on his blog do have alt text, but it's just the LaTeX code, which isn't much better.) LLMs are extremely good at outputting LaTeX, ChatGPT will output LaTeX, which the website will render as such. Why do you think LLMs have trouble understanding it? | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
▲ | MengerSponge 2 days ago | parent | prev [-] | |||||||||||||||||||||||||
LLMs are decent with LaTeX! It's just markup code after all. I've heard from some colleagues that they can do decent image to code conversion for a picture of an equation or even some handwritten ones. |