What it is and why it matters
Generative AI consumes existing data, learns from it, then generates data with similar characteristics. For example, it can generate text, images, audio, video and computer code.
The evolution of generative AI
Traditional AI and machine learning systems recognize patterns in data to make predictions. But generative AI goes beyond predicting – it generates new data as its primary output. Imagine receiving the full text for a speech just seconds after giving a chatbot (like ChatGPT) a few words to describe your idea. Generating music, art or images from text-based descriptions. Or developing a business strategy through conversational, back-and-forth “prompting” with a generative AI tool.
Where did it all start?
Contrary to popular opinion, generative AI is not new – it’s built from technologies we’ve used for decades, including AI, machine learning and statistical methods. Three core generative AI technologies are digital twins, large language models and synthetic data generation.
While the origins of generative AI could be traced farther back, we’ll start with 1966 and a chatbot named ELIZA.
Joseph Weizenbaum, who built ELIZA, designed it to imitate Rogerian psychotherapists who mirror what the patient says. ELIZA used pattern matching to accomplish this feat. ELIZA was one of the first programs to attempt the Turing Test – an imitation game that tests a machine’s ability to exhibit intelligent behavior like a human.
As methods of analyzing unstructured text data evolved, the 1970s through the 1990s saw growth in semantic networks, ontologies, recurrent neural networks and more. From 2000 through 2015, language modeling and word embedders improved, and Google Translate emerged.
In 2014, Ian Goodfellow and colleagues developed the generative adversarial network (GAN), setting up two neural networks to compete (i.e., train) against each other. One network generated data while the other tried to determine if the data was real or fake. Transformer models were introduced in 2017. They included a self-attention mechanism that allowed them to weigh the importance of different parts of the input when making predictions. Architectures like BERT and ELMo became popular, too.
Generative pre-trained transformer (GPT) models appeared next, with the first GPT model arriving in 2018. This model was trained on large quantities of text data from the internet. With 117 million parameters, it could generate text similar in style and content to the training data. By 2023, large language GPT models had evolved to a point where they could perform proficiently on difficult exams, like the bar exam.
The rapid rise of generative AI tech
A disruptive technology, the impact of generative AI has been compared to discoveries like electricity and the printing press. With the potential to drastically boost productivity, conversational AI models like ChatGPT have rocketed in popularity among business and everyday users – and raised concerns about data privacy, bias in AI, ethics and accuracy. The global market for generative AI is expected to grow to $110.8 billion by 2030.
Who's using generative AI?
Generative AI spans a wide range of industries and business functions across the world. As it grows in popularity, the technology has simultaneously triggered excitement and fear among individuals, businesses and government entities. Let’s look at how some industries are using generative AI today.
The results of generative AI, at their core, are a reflection of us, humans. ... Consumers must continue to apply critical thinking whenever interacting with conversational AI and avoid automation bias (the belief that a technical system is more likely to be accurate and true than a human). Reggie Townsend VP of the SAS Data Ethics Practice
Considerations for generative AI models
Models are expensive to run – requiring tremendous amounts of compute power and data. You should carefully evaluate the ROI before implementing a generative AI model. There are ethical considerations, too. Where did the data come from – and who owns it? Is it trustworthy? Do you understand precisely how the model was built?
How generative AI works
Some popular examples of generative AI technologies include DALL-E, an image generation system that creates images from text inputs, ChatGPT (a text generation system), the Google Bard chatbot and Microsoft's AI-powered Bing search engine. Another example is using generative AI to create a digital representation of a system, business process or even a person – like a dynamic representation of someone’s current and future health status.
There are three main types of generative technologies (digital twins, large language models and synthetic data generation).
Digital twins are virtual models of real-life objects or systems built from data that is historical, real-world, synthetic or from a system’s feedback loop. They’re built with software, data, and collections of generative and non-generative models that mirror and synchronize with a physical system – such as an entity, process, system or product. Digital twins are used to test, optimize, monitor or predict. For example, a digital twin of a supply chain can help companies predict when shortages may occur.
A large language model (LLM) is a powerful machine learning model that can process and identify complex relationships in natural language, generate text and have conversations with users. These models rely on techniques like deep learning and neural networks. Defined as natural language-processing AI models, LLMs are trained on massive amounts of text data. The resulting models have up to billions of parameters. OpenAI’s ChatGPT is an example of a popular large language model.
Synthetic data generation refers to on-demand, self-service or automated data generated by algorithms or rules rather than collected from the real world. Synthetic data is often generated to meet conditions lacking in real data. It reproduces the same statistical properties, probabilities, patterns and characteristics as the real-world data from which it’s trained. Many organizations use synthetic data to preserve privacy or to overcome other challenges with collecting and using real-world data, such as cost, time-intensive data preparation processes or bias.
Many other technologies enable and support generative AI:
An algorithm is a list of step-by-step instructions designed to accomplish a specific task or solve a problem. Many computer programs are a sequence of algorithms written in a way the computer can understand. As algorithms begin to supplement or replace human decisions, we must explore their fairness and demand transparency into how they’re developed.
Artificial intelligence makes it possible for machines to learn from experience, adjust to new inputs and perform humanlike tasks. AI often relies heavily on deep learning and NLP. Through such technologies, computers can be trained to accomplish specific tasks by processing large amounts of data and recognizing patterns.
Deep learning is a subset of machine learning that trains a computer to perform humanlike tasks, such as recognizing speech, identifying images and making predictions. It improves the ability to classify, recognize, detect and describe using data. Deep learning models like GANs and variational autoencoders (VAEs) are trained on massive data sets and can generate high-quality data. Newer techniques like StyleGANs and transformer models can create realistic videos, images, text and speech.
Machine learning is a method of data analysis that automates analytical model building. It’s a branch of artificial intelligence that trains a machine how to learn. Machine learning is based on the idea that systems can learn from data, identify patterns and make decisions with minimal human intervention.
Natural language processing is a branch of artificial intelligence that helps computers understand, interpret and manipulate human language. NLP draws from many disciplines, including computer science and computational linguistics, to fill the gap between human communication and computer understanding.
Neural networks are computing systems with interconnected nodes that work much like neurons in the human brain. Neural networks use algorithms to recognize hidden patterns and correlations in raw data, cluster and classify it, and continuously learn and improve over time.
Reinforcement learning is when an algorithm discovers through trial and error which actions produce the greatest rewards. A machine learning model, reinforcement learning relies on a reward signal for its feedback mechanism as it gradually learns the best (or most rewarding) policy or goal. It’s often used for robotics, gaming and navigation.
5 steps for fine-tuning a model
Generative AI relies on many different AI algorithms and technologies to generate data that has similar probabilistic distributions and characteristics to the data from which it learns. Rather than building from scratch, you can follow these five steps to fine-tune a pre-trained foundational large language model.
1. Define the task.
Choose a suitable pre-trained large language model and clearly define the task for which it’s being fine-tuned. This could be text classification (namely entity recognition), text generation, etc.
2. Prepare the data.
Gather and preprocess your task-specific data – for tasks like labeling, formatting and tokenization. Create training and validation (and possibly test) data sets.
Train the modified model on your task-specific data, using the training data set to update the model's weight. Monitor the model's performance on the validation set to prevent overfitting.
4. Evaluate and test.
After training, evaluate your fine-tuned model on the validation set, making necessary adjustments based on results. When satisfied, test the model on the test set to get an unbiased estimate of performance.
When you're confident in the model's performance, deploy it for its intended use. This could involve integrating the model into an application, website or another platform.
What is synthetic data?
Data is essential for building models, but high-quality data can be hard to find, biased or expensive. One way to solve those issues is by using synthetic data, which is created artificially (often with algorithms). If we use real-world data sets to generate additional, synthetic data – with appropriate properties for building good machine learning models – we can train models for virtually any purpose, like researching a rare disease.
An AI and analytics platform
With SAS® Viya®, there’s no such thing as too much information. Learn about the quickest way to get from a billion points of data to a point of view.
- Article 4 strategies that will change your approach to fraud detectionAs fraudulent activity grows and fighting fraud becomes more costly, financial institutions are turning to anti-fraud technology to build better arsenals for fraud detection. Discover four ways to improve your organization's risk posture. 23 May 2023
- Article The untapped potential in unstructured textText is the largest human-generated data source. It grows every day as we post on social media, interact with chatbots and digital assistants, send emails, conduct business online, generate reports and essentially document our daily thoughts and activities using computers and mobile devices. 19 June 2019
- Article What are chatbots?Chatbots are a form of conversational AI designed to simplify human interaction with computers. Learn how chatbots are used in business and how they can be incorporated into analytics applications. 21 February 2019
- Article How to improve your AI marketing skillsMarketing teams can use current AI capabilities to enhance their efforts around campaign automation, dynamic pricing based on forecasting models, and by providing more relevant, real-time customer offers. 19 February 2018