Beyond Large Language Models: Seeking Alternative Pathways for AI Advancement

Technology News

Beyond Large Language Models: Seeking Alternative Pathways for AI Advancement
Artificial IntelligenceLarge Language ModelsAI Development
  • 📰 ForbesTech
  • ⏱ Reading Time:
  • 650 sec. here
  • 17 min. at publisher
  • 📊 Quality Score:
  • News: 280%
  • Publisher: 59%

The current focus on large language models (LLMs) and generative AI may be misguided, given their limitations and potential negative consequences. Critics suggest exploring alternative AI approaches, as LLMs haven't delivered the promised societal transformation. The article highlights concerns about misinformation, job displacement, and the challenges of explainable AI.

It’s time to move past large language models and create a new narrative. The hiccups we’ve experienced from large language models and generative AI — a still-novel technology — since its inception a few years ago is evidence that this may not be the pathway towards the promised intelligence that will transform society, as we have been led to believe.

There have been more ebbs in the journey of large language models than significant milestones achieved. This, among top frontier tech companies, where the billions in investment have been concentrated.against frontier models, these are the events defining this technology. Righteous defenders, in lock step, are building the solutions toin the process. This is not only adding to the fear of widespread fakes, but validating the incessant fears of job loss, and aWhat are we missing? Perhaps, it’s time to move past LLMs and surface alternatives. What are the pathways that LLMs, alone, cannot muster?, that in his development of deep learning to mimic human decision making, the need for regulators to mandate explanations for AI system decisions would be a "complete disaster" because humans themselves cannot fully explain their own decision-making processes. He explained, “People can’t explain how they work, for most of the things they do. When you hire somebody, the decision is based on all sorts of things you can quantify, and then all sorts of gut feelings. People have no idea how they do that. If you ask them to explain their decision, you are forcing them to make up a story.”Gary Marcus is a professor emeritus at New York University, cognitive scientist, and author of The Algebraic Mind, and Rebooting AI. He has been a renowned critic of the prevailing large language models for some time. He recognized, in his latestwith the NY Times, “GPT-5 is a step forward but nowhere near the A.I. revolution many had expected. That is bad news for the companies and investors who placed substantial bets on the technology.” The times have changed, and he admits that since 2019 when he started criticizing the industry focus on LLMs and the underlying deep learning technology, few were willing to consider anything else. In response to Hinton’s argument dismissing the need for explainability, Marcus remarks, “He made a variation of that argument the other day, and then Nobel Prize Committee actually retweeted it. It was embarrassing. It was a slightly different argument but structured similarly. It's like, well, humans hallucinate, so the fact that LLMs hallucinate means they're like us and it's all fine.” He dismisses the goal of AI is to replicate human intelligence, remarking, “Humans still do a bunch of things that machines don’t do very well, such as learning new skills, and reasoning with abstract ideas. We have a fluidity to our thought, a flexibility to our thought that AI still lacks. We can reasonably hope that we might learn something from people that might allow our AI to work better, even if we’re not building replicas.” He explains, “The history of AI, in the early days, is where people took cognitive science seriously. In the last several years, the field of AI has been taken over by people who do a form of statistical learning that doesn’t owe that much to what humans do. The leaders of the field have been dismissive of almost anything other than a small set of techniques. And they did this with the hypothesis that scaling would get to AGI. And I don’t think it has. And so, we should go back to thinking about what cognitive science can teach us.” Mounir Shita is a 30-year veteran AGI researcher and founder of EraNova Global, a physics-first program of work on the nature of intelligence. His“models a goal as a physically realizable future state.” He defines intelligence as the capacity to steer “causal chains toward the state under the constraints of physics.” In 2010 Shita cofounded Kimera Systems to implement these ideas in production. He rejects the notion that understanding human behavior is the holy grail to AGI. His analogy of the Wright brothers centers on the mechanisms that enabled birds to fly: “They didn’t build a flapping machine; they liberated flight from birds by identifying the laws—lift, drag, thrust, control—and then engineering within those laws. That’s why we fly thousands of passengers and tons of cargo. Laws scale. Likeness doesn’t,” Shita explains. Marc Fawzi is a systems thinker, where he built the first commercial low-code platform for telecom service providers. Early startup work with MIT’s AI Lab in the late 1980’s exposed him to neural-network research and massively parallel compute architectures.is where cognition relies on historical events that can encode the foundation but will update with new evidence. The second layer:identifies the similarities and constraints within a system of concepts — that will ultimately preserve that meaning and frame. For example, Surrealism is an art and literally movement and its primary constraint is to reject rational and traditional conventions, so employs artistic techniques to support this frame. The third layer:is how the introduction of something can be explained from represented systems Finally, the fourth layer: the objectives, will define the preferences, risks and tradeoffs. Fawzi emphasizes, “All four layers must cohere. If the structure fails to constrain statistics or inference or ignore structure, you get fluent reasoning that drifts from truth; if objectives are unclear, you get mis-aimed competence.” In other words, intelligence is the alignment of uncertainty, structure, inference and objectives. Shita stated the definition of intelligence is similar to Fawzi, and within the structure of the universe: the state, goals, dynamics, time, causality. He affirms, “Derive the update/control laws, state their regime of validity, then build. That’s how science and engineering have advanced for millennia... If a system can’t adapt mid-episode when the rules flip — and explain the causal hypothesis behind its update — you’re not building AGI. You’re staging a demo.””, suspects venture capitalists who have poured billions into big frontier models, prefer a simple story that justifies a big bet which is, “if we keep making these things bigger, we’ll get artificial general intelligence.” He says VC investment in LLMs is the belief that more data and more compute are key. “Even if they don’t make any money on the investment, they get 2% each year of what’s under management and that incentivizes them to have plausible stories for a significant upside. And so, to the investors, it sounds plausible that maybe we’ll get to AGI by just doing the scaling.”The excitement was in the belief that more compute would produce smarter models. Marcus said that people took these “alleged” laws seriously projecting how much compute we need and that will get us to AGI, “and I think what a lot of people saw recently was GPT-5 was supposed us to take us a long way there, and it didn’t take us nearly as far as people expected.”increasing costs are often used as a yardstick for saying a problem is intractable,” or uncontrollable.that by 2030, “data centers are projected to require $6.7 trillion worldwide to keep pace with the demand for compute power.” The capital expenditure alone is projected to be $5.2 trillion." Data Centers will now compete with cities for energy. The demand for compute is largely uncertain, hence risky, and directly impacts how much capital to allocate building these massive and expensive infrastructures.According to Shita, language belongs in the core knowledge base of AGI — as learned knowledge. However, LLMs don’t. He explains, “LLMs are trained to maximize next-token likelihood on offline text. They model words about the world, not the world those words must change. Without live sensors, interventions and updates to model and policy from consequences, their beliefs stay anchored to the distribution of past language, which is decoupled from present state and outcomes.” Shita describes AGI as a ‘causal instrument’ that should change minds, policies, markets and machines toward a goal. However, that requires tying language to be a goal-conditioned model of the environment, subject to constraints and feedback. For Shita, LLMs fall under Fawzi’s first layer of statistics — “they make excellent priors and interfaces but are not a core controller., a collaboration, and ethical AI commons to train an AI, rooted in justice, and accountability, agrees with Shita. LLMs, by themselves are not the answer, and adds, “With grounding , interaction and verifiers, they can participate in layer alignment.” However, without ties to constraints he suggests that language fluence does not equal knowledge. “Patterns of words aren’t automatically patterns of the world.” Marcus says the field abdicated its duty to consider better ideas for what to do with the data. He says, they started running with LLMs... and there were many problems with LLMs that were foreseeable, and they spoke about “emergence a lot, and that whatever was necessary would emerge almost as if by magic if it was enough data. Instead, what we see, from GPT-5, Grok4 and Llama4 is that while you get some value by adding more data, we’re really in a regime where more data alone is giving us diminishing returns.” Marcus concludes, “This whole paradigm is basically a souped-up regurgitation. And not everything you need is in the training set. And when it’s not in the training set or it's dissimilar in an important way from the training data, this paradigm does not work that well... especially in the hard cases.”So, throwing more data and more compute doesn’t necessarily bring us closer to AGI and the AI community is experiencing a reckoning that deep neural networks are not the panacea they made them out to be. Marcus points to this retreat from LLMs and towards “good old fashioned symbol manipulating devices like code interpreters like Python, JavaScript.” He recently wrote about howfor the inclusion of algebraic systems , “systems for explicitly representing structured representations” and “database-like systems to distinguish individuals from kinds.” For the latter, he forewarned that in their absence, “hallucinations would emerge as a form of overgeneralization.” Shita, he sees things differently. He insists, time and causality are also critical to intelligence. A goal, he states, is tied to a future state of the world we want to achieve. So, they are fundamentally tied to time. He explains, “To change anything in the world, you must understand causality—the laws of cause and effect. The world is always in motion. Intelligence is the ability to skillfully intervene in these ongoing processes, redirecting them to achieve a specific outcome.” Dynamics acknowledges that the world is constantly changing, which is why we AI can't rely on a static, pre-trained model. It must use a dynamic model that continuously updates its understanding of reality for accurate predictions. He references Judea Pearl , who formalized causal reasoning using structured models, as well as, whose research revealed the importance of temporal memory and how the brain’s view of the world, as episodes relative to each other: where it records sensations but also what happened, where and when. Fawzi emphasizes that one of the biggest gaps evident in today’s environment is the lack of one world model, “priors that hold across domains and time, calibrated and updatable under feedback.”, father of Reinforcement Learning, recently said “He says that LLMs have no model of the world, and have no goals... Large Language Models are about mimicking people, doing what people say you should do. They’re not about figuring out what to do.” Sutton, andGary Marcus says, at least for now, we should give up on AGI, stating, “In the long term, AGI will be, or has a chance of being net helpful to humanity. But right now, we don’t actually know how to build it. The technologies we have now really struggle with reliability. They can’t follow instructions. It makes it difficult to preclude them from leading people into delusions, or talking them into suicide or making biological weapons.” He explains it is not well controlled and, in many ways, it’s not that good. He advocates for going back to the ‘good old days’ and focusing on dedicated machines that work on dedicated problems. He references Mounir Shita claims that computers are 100% deterministic, “given inputs, code, and state, the next state is fixed. Free will cannot emerge from deterministic machinery. The real question isn’t can it rebel, but who raises it, what it learns to value and how it reasons when it comes to consequences for others.” He expresses that AGI raised “narrowly — socialized to pursue individual interests while discounting others — will optimize that bias with unprecedented efficiency.” He argues “we need to stop treating AGI as something that needs guardrails. Bolted on guardrails shrink an agent’s decision space and make it dumber, not safer; Embedded ethics, however, expand competence.” For now, with many of the AI research community aligned that the current prevailing LLM technology is not IT, time will tell whether the massive investments to secure its future, will also fold.

We have summarized this news so that you can read it quickly. If you are interested in the news, you can read the full text here. Read more:

ForbesTech /  🏆 318. in US

Artificial Intelligence Large Language Models AI Development Generative AI AI Ethics

 

United States Latest News, United States Headlines

Similar News:You can also read news stories similar to this one that we have collected from other news sources.

Data Pipelines And Prompt Engineering Matter More For RAG Than LLMsData Pipelines And Prompt Engineering Matter More For RAG Than LLMsDon’t chase the newest model without first ensuring your pipelines and prompts can support it.
Read more »

Congressional leaders leave White House with no deal to avert shutdown: ‘Still large differences’Congressional leaders leave White House with no deal to avert shutdown: ‘Still large differences’Democrats and Republicans are no closer to a deal to avert a government shutdown, with the funding deadline looming in just over 24 hours.
Read more »

Democrats and Republicans face 'large differences' ahead of shutdown deadlineDemocrats and Republicans face 'large differences' ahead of shutdown deadlineAfter canceling an earlier meeting, President Trump met with a bipartisan group of congressional leaders on Monday. Lawmakers have until the end of the day on Sept. 30 to avoid a government shutdown.
Read more »

New Yorkers fight against large e-battery storage plants | 7 On Your Side InvestigatesNew Yorkers fight against large e-battery storage plants | 7 On Your Side InvestigatesWe've seen how an e-battery can catch fire and explode in a matter of seconds and that's from a single battery on an electric bike or scooter.
Read more »

Precision Mental Health Gets Precisely Boosted Via Innovative Uses Of Advanced AI And LLMsPrecision Mental Health Gets Precisely Boosted Via Innovative Uses Of Advanced AI And LLMsStanford presented the latest research in precision mental health and the use of AI. I share some key insights on the emerging topic. It's an AI Insider scoop.
Read more »

USD: A small piece of a large puzzleUSD: A small piece of a large puzzleThe US Dollar (USD) is set to continue weakening in the coming months. However, what is particularly interesting are the explanations other analysts offer for their respective forecasts. Recently, one argument has surfaced repeatedly: the increasing demand for hedging against USD weakness.
Read more »



Render Time: 2026-04-01 19:21:08