Apple’s study proves that LLM-based AI models are flawed because they cannot reason

A new paper from Apple’s artificial intelligence scientists has found that engines based on large language models, such as those from Meta and OpenAI, still lack basic reasoning skills.

The group has proposed a new benchmark, GSM-Symbolic, to help others measure the reasoning capabilities of various large language models (LLMs). Their initial testing reveals that slight changes in the wording of queries can result in significantly different answers, undermining the reliability of the models.

Read more: appleinsider.com


Comments

Leave a Reply

Your email address will not be published. Required fields are marked *