Language AIs in 2024: Size, guardrails and steps toward AI agents
- Written by John Licato, Associate Professor of Computer Science, Director of AMHR Lab, University of South Florida
I research[1] the intersection of artificial intelligence, natural language processing and human reasoning as the director of the Advancing Human and Machine Reasoning lab[2] at the University of South Florida. I am also commercializing this research in an AI startup[3] that provides a vulnerability scanner for language models.
From my vantage point, I observed significant developments in the field of AI language models in 2024, both in research and the industry.
Perhaps the most exciting of these are the capabilities of smaller language models, support for addressing AI hallucination, and frameworks for developing AI agents[4].
Small AIs make a splash
At the heart of commercially available generative AI products like ChatGPT are large language models, or LLMs, which are trained on vast amounts of text and produce convincing humanlike language. Their size is generally measured in parameters[5], which are the numerical values a model derives from its training data. The larger models like those from the major AI companies have hundreds of billions of parameters.
There is an iterative interaction between large language models and smaller language models[6], which seems to have accelerated in 2024.
First, organizations with the most computational resources experiment with and train increasingly larger and more powerful language models. Those yield new large language model capabilities, benchmarks, training sets and training or prompting tricks. In turn, those are used to make smaller language models – in the range of 3 billion parameters or less – which can be run on more affordable computer setups, require less energy and memory to train, and can be fine-tuned with less data.
No surprise, then, that developers have released a host of powerful smaller language models – although the definition of small keeps changing: Phi-3[7] and Phi-4[8] from Microsoft, Llama-3.2 1B and 3B[9], and Qwen2-VL-2B[10] are just a few examples.
These smaller language models can be specialized for more specific tasks, such as rapidly summarizing a set of comments or fact-checking text against a specific reference. They can work with their larger cousins[11] to produce increasingly powerful hybrid systems.
Wider access
Increased access to highly capable language models large and small can be a mixed blessing. As there were many consequential elections around the world in 2024, the temptation for the misuse of language models was high.
Language models can give malicious users the ability to generate social media posts and deceptively influence public opinion. There was a great deal of concern[12] about this threat in 2024, given that it was an election year in many countries.
And indeed, a robocall faking President Joe Biden’s voice asked New Hampshire Democratic primary voters to stay home[13]. OpenAI had to intervene to disrupt over 20 operations and deceptive networks[14] that tried to use its models for deceptive campaigns. Fake videos and memes were created and shared[15] with the help of AI tools.
Despite the anxiety surrounding AI disinformation[16], it is not yet clear what effect these efforts actually had[17] on public opinion and the U.S. election. Nevertheless, U.S. states passed a large amount of legislation in 2024[18] governing the use of AI in elections and campaigns.
Misbehaving bots
Google started including AI overviews[19] in its search results, yielding some results that were hilariously and obviously wrong – unless you enjoy glue in your pizza[20]. However, other results may have been dangerously wrong, such as when it suggested mixing bleach and vinegar[21] to clean your clothes.
Large language models, as they are most commonly implemented, are prone to hallucinations[22]. This means that they can state things that are false or misleading, often with confident language. Even though I and others[23] continually beat the drum about this, 2024 still saw many organizations learning about the dangers of AI hallucination the hard way.
Despite significant testing, a chatbot playing the role of a Catholic priest advocated for baptism via Gatorade[24]. A chatbot advising on New York City laws and regulations[25] incorrectly said it was “legal for an employer to fire a worker who complains about sexual harassment, doesn’t disclose a pregnancy or refuses to cut their dreadlocks.” And OpenAI’s speech-capable model forgot whose turn it was to speak and responded to a human in her own voice[26].
Fortunately, 2024 also saw new ways to mitigate and live with AI hallucinations. Companies and researchers are developing tools for making sure AI systems follow given rules pre-deployment[27], as well as environments to evaluate them[28]. So-called guardrail frameworks[29] inspect large language model inputs and outputs in real time, albeit often by using another layer of large language models.
And the conversation on AI regulation accelerated[30], causing the big players in the large language model space to update their policies on responsibly scaling[31] and harnessing AI[32].
But although researchers are continually finding ways to reduce hallucinations[33], in 2024, research convincingly showed[34] that AI hallucinations are always going to exist in some form[35]. It may be a fundamental feature of what happens when an entity has finite computational and information resources. After all, even human beings are known to confidently misremember and state falsehoods[36] from time to time.
The rise of agents
Large language models, particularly those powered by variants of the transformer architecture[37], are still driving the most significant advances in AI. For example, developers are using large language models to not only create chatbots, but to serve as the basis of AI agents. The term “agentic AI” shot to prominence in 2024[38], with some pundits even calling it the third wave[39] of AI.
To understand what an AI agent[40] is, think of a chatbot expanded in two ways: First, give it access to tools that provide the ability to take actions[41]. This might be the ability to query an external search engine, book a flight or use a calculator. Second, give it increased autonomy, or the ability to make more decisions on its own.
For example, a travel AI chatbot might be able to perform a search of flights based on what information you give it, but a tool-equipped travel agent might plan out an entire trip itinerary, including finding events, booking reservations and adding them to your calendar.
In 2024, new frameworks for developing AI agents emerged. Just to name a few, LangGraph[42], CrewAI[43], PhiData[44] and AutoGen/Magentic-One[45] were released or improved in 2024.
Companies are just beginning to adopt[46] AI agents. Frameworks for developing AI agents are new and rapidly evolving. Furthermore, security, privacy and hallucination risks are still a concern.
But global market analysts forecast this to change[47]: 82% of organizations surveyed plan to use agents within 1-3 years[48], and 25% of all companies currently using generative AI[49] are likely to adopt AI agents in 2025.
References
- ^ I research (scholar.google.com)
- ^ Advancing Human and Machine Reasoning lab (github.com)
- ^ AI startup (www.actualization.ai)
- ^ AI agents (theconversation.com)
- ^ parameters (www.thecloudgirl.dev)
- ^ large language models and smaller language models (www.youtube.com)
- ^ Phi-3 (news.microsoft.com)
- ^ Phi-4 (techcommunity.microsoft.com)
- ^ Llama-3.2 1B and 3B (huggingface.co)
- ^ Qwen2-VL-2B (huggingface.co)
- ^ work with their larger cousins (aclanthology.org)
- ^ great deal of concern (campaignlegal.org)
- ^ to stay home (www.nbcnews.com)
- ^ disrupt over 20 operations and deceptive networks (openai.com)
- ^ created and shared (www.npr.org)
- ^ anxiety surrounding AI disinformation (washingtonstatestandard.com)
- ^ not yet clear what effect these efforts actually had (time.com)
- ^ legislation in 2024 (www.ncsl.org)
- ^ AI overviews (blog.google)
- ^ glue in your pizza (www.forbes.com)
- ^ mixing bleach and vinegar (www.salon.com)
- ^ prone to hallucinations (doi.org)
- ^ others (www.youtube.com)
- ^ advocated for baptism via Gatorade (www.businessinsider.com)
- ^ advising on New York City laws and regulations (apnews.com)
- ^ responded to a human in her own voice (arstechnica.com)
- ^ follow given rules pre-deployment (doi.org)
- ^ environments to evaluate them (doi.org)
- ^ guardrail frameworks (techcrunch.com)
- ^ on AI regulation accelerated (www.ncsl.org)
- ^ responsibly scaling (www.anthropic.com)
- ^ harnessing AI (openai.com)
- ^ ways to reduce hallucinations (doi.org)
- ^ convincingly showed (doi.org)
- ^ hallucinations are always going to exist in some form (doi.org)
- ^ confidently misremember and state falsehoods (health.clevelandclinic.org)
- ^ transformer architecture (dl.acm.org)
- ^ shot to prominence in 2024 (trends.google.com)
- ^ third wave (www.forbes.com)
- ^ AI agent (theconversation.com)
- ^ ability to take actions (python.langchain.com)
- ^ LangGraph (www.langchain.com)
- ^ CrewAI (www.crewai.com)
- ^ PhiData (www.phidata.com)
- ^ AutoGen/Magentic-One (www.microsoft.com)
- ^ beginning to adopt (www.forbes.com)
- ^ forecast this to change (www.analyticsvidhya.com)
- ^ plan to use agents within 1-3 years (www.capgemini.com)
- ^ 25% of all companies currently using generative AI (www2.deloitte.com)
Authors: John Licato, Associate Professor of Computer Science, Director of AMHR Lab, University of South Florida
Read more https://theconversation.com/language-ais-in-2024-size-guardrails-and-steps-toward-ai-agents-245646