Search
Items tagged with: LLMs
Here is an example of how bad #LLMs are with math. I asked about velocity in the context of Agile process. The answer?
“Sure! Let's say that an Agile development team has completed four iterations, each lasting two weeks. In the first iteration, they delivered 12 user stories; in the second, they delivered 10; in the third, they delivered 9; and in the fourth, they delivered 8. The total number of user stories completed by the end of the fourth iteration is 49 (12 + 10 + 9 + 8)”.
“Sure! Let's say that an Agile development team has completed four iterations, each lasting two weeks. In the first iteration, they delivered 12 user stories; in the second, they delivered 10; in the third, they delivered 9; and in the fourth, they delivered 8. The total number of user stories completed by the end of the fourth iteration is 49 (12 + 10 + 9 + 8)”.
Let’s be honest, if you’re a software engineer, you know where all this compute and power consumption is going. While it’s popular to blame #LLMs, y’all know how much is wasted on #docker, microservices, overscaled #kubernetes, spark/databricks and other unnecessary big data tech. It’s long past time we’re honest with the public about how much our practices are hurting the climate, and stop looking for scapegoats https://thereader.mitpress.mit.edu/the-staggering-ecological-impacts-of-computation-and-the-cloud/
The Staggering Ecological Impacts of Computation and the Cloud
Anthropologist Steven Gonzalez Monserrate draws on five years of research and ethnographic fieldwork in server farms to illustrate some of the diverse environmental impacts of data storage.The MIT Press Reader
This is a really nice read about whether #LLMs can actually reason: https://aiguide.substack.com/p/can-large-language-models-reason
I think expecting language models to reason like the math engines maght be a bit out of range! Nice try!
I think expecting language models to reason like the math engines maght be a bit out of range! Nice try!
Can Large Language Models Reason?
What should we believe about the reasoning abilities of today’s large language models? As the headlines above illustrate, there’s a debate raging over whether these enormous pre-trained neural networks have achieved humanlike reasoning abilities, or …Melanie Mitchell (AI: A Guide for Thinking Humans)
Three years in the making - our big review/opinion piece on the capabilities of large language models (LLMs) from the cognitive science perspective.
Thread below! 1/
#AI #cogneuro #NLP #LLMs #languageandthought
https://arxiv.org/abs/2301.06627
Thread below! 1/
#AI #cogneuro #NLP #LLMs #languageandthought
https://arxiv.org/abs/2301.06627
Dissociating language and thought in large language models: a cognitive perspective
Today's large language models (LLMs) routinely generate coherent, grammatical and seemingly meaningful paragraphs of text.arXiv.org