Remix.run Logo
TinyLoRA – Learning to Reason in 13 Parameters(arxiv.org)
68 points by sorenjan 5 days ago | 6 comments
a-t-c-g 23 minutes ago | parent | next [-]

The quality of custom models trained with proper reasoning datasets[0] even with small parameters (3-7B is sweet spot) is incredible now

[0]: cartesien.io or Salesforce's WebscaleRL

measurablefunc an hour ago | parent | prev [-]

With four parameters I can fit an elephant, and with five I can make him wiggle his trunk so there is still room for improvement.

esafak an hour ago | parent [-]

Except learning to reason is a far cry from curve fitting. Our brains have more than five parameters.

voxelghost 12 minutes ago | parent | next [-]

After a quick content browse, my understanding is this is more like with a very compressed diff vector, applied to a multi billion parameter model, the models could be 'retrained' to reason (score) better on a specific topic , e.g. math was used in the paper

ekuck 14 minutes ago | parent | prev | next [-]

speak for yourself!

est 30 minutes ago | parent | prev [-]

reasoning capability might just be some specific combinations of mirror neurons.

even some advanced math usually evolves applying patterns found elsewhere into new topics