Natural Language Processing (NLP)
What it is and why it matters
Natural language processing (NLP) is a branch of artificial intelligence that helps computers understand, interpret and manipulate human language. NLP draws from many disciplines, including computer science and computational linguistics, in its pursuit to fill the gap between human communication and computer understanding.
Evolution of natural language processing
While natural language processing isn’t a new science, the technology is rapidly advancing thanks to an increased interest in human-to-machine communications, plus an availability of big data, powerful computing and enhanced algorithms.
As a human, you may speak and write in English, Spanish or Chinese. But a computer’s native language – known as machine code or machine language – is largely incomprehensible to most people. At your device’s lowest levels, communication occurs not with words but through millions of zeros and ones that produce logical actions.
Indeed, programmers used punch cards to communicate with the first computers 70 years ago. This manual and arduous process was understood by a relatively small number of people. Now you can say, “Alexa, I like this song,” and a device playing music in your home will lower the volume and reply, “OK. Rating saved,” in a humanlike voice. Then it adapts its algorithm to play that song – and others like it – the next time you listen to that music station.
Let’s take a closer look at that interaction. Your device activated when it heard you speak, understood the unspoken intent in the comment, executed an action and provided feedback in a well-formed English sentence, all in the space of about five seconds. The complete interaction was made possible by NLP, along with other AI elements such as machine learning and deep learning.
Make Every Voice Heard with Natural Language Processing
Discover how machines can learn to understand and interpret the nuances of human language; how AI, natural language processing and human expertise work together to help humans and machines communicate and find meaning in data; and how NLP is being used in multiple industries.
Community outreach and support for COPD patients enhanced through natural language processing and machine learning
The COPD Foundation uses text analytics and sentiment analysis, NLP techniques, to turn unstructured data into valuable insights. These findings help provide health resources and emotional support for patients and caregivers. Learn more about how analytics is improving the quality of life for those living with pulmonary disease.
Why is NLP important?
Large volumes of textual data
Natural language processing helps computers communicate with humans in their own language and scales other language-related tasks. For example, NLP makes it possible for computers to read text, hear speech, interpret it, measure sentiment and determine which parts are important.
Today’s machines can analyze more language-based data than humans, without fatigue and in a consistent, unbiased way. Considering the staggering amount of unstructured data that’s generated every day, from medical records to social media, automation will be critical to fully analyze text and speech data efficiently.
Structuring a highly unstructured data source
Human language is astoundingly complex and diverse. We express ourselves in infinite ways, both verbally and in writing. Not only are there hundreds of languages and dialects, but within each language is a unique set of grammar and syntax rules, terms and slang. When we write, we often misspell or abbreviate words, or omit punctuation. When we speak, we have regional accents, and we mumble, stutter and borrow terms from other languages.
While supervised and unsupervised learning, and specifically deep learning, are now widely used for modeling human language, there’s also a need for syntactic and semantic understanding and domain expertise that are not necessarily present in these machine learning approaches. NLP is important because it helps resolve ambiguity in language and adds useful numeric structure to the data for many downstream applications, such as speech recognition or text analytics.
NLP in today’s world
Learn more about natural language processing in many industries
Planning for NLP
How are organizations around the world using artificial intelligence and NLP? What are the adoption rates and future plans for these technologies? What are the budgets and deployment plans? And what business problems are being solved with NLP algorithms? Find out in this report from TDWI.
Natural language processing for government efficiency
Government agencies are bombarded with text-based data, including digital and paper documents. Using technologies like natural language processing, text analytics and machine learning, agencies can reduce cumbersome, manual processes while addressing citizen demands for transparency and responsiveness, solving workforce challenges and unleashing new insights from their data.
What can text analytics do for your organization?
Text analytics is a type of natural language processing that turns text into data for analysis. Learn how organizations in banking, health care and life sciences, manufacturing and government are using text analytics to drive better customer experiences, reduce fraud and improve society.
How does NLP work?
Breaking down the elemental pieces of language
Natural language processing includes many different techniques for interpreting human language, ranging from statistical and machine learning methods to rules-based and algorithmic approaches. We need a broad array of approaches because the text- and voice-based data varies widely, as do the practical applications.
Basic NLP tasks include tokenization and parsing, lemmatization/stemming, part-of-speech tagging, language detection and identification of semantic relationships. If you ever diagramed sentences in grade school, you’ve done these tasks manually before.
In general terms, NLP tasks break down language into shorter, elemental pieces, try to understand relationships between the pieces and explore how the pieces work together to create meaning.
These underlying tasks are often used in higher-level NLP capabilities, such as:
- Content categorization. A linguistic-based document summary, including search and indexing, content alerts and duplication detection.
- Large Language Model (LLM)-based classification. BERT-based classification is used to capture the context and meaning of words in a text to improve accuracy compared to traditional models.
- Corpus Analysis. Understand corpus and document structure through output statistics for tasks such as sampling effectively, preparing data as input for further models and strategizing modeling approaches.
- Contextual extraction. Automatically pull structured information from text-based sources.
- Sentiment analysis. Identifying the mood or subjective opinions within large amounts of text, including average sentiment and opinion mining.
- Speech-to-text and text-to-speech conversion. Transforming voice commands into written text, and vice versa.
- Document summarization.Automatically generating synopses of large bodies of text and detect represented languages in multi-lingual corpora (documents).
- Machine translation. Automatic translation of text or speech from one language to another.
In all these cases, the overarching goal is to take raw language input and use linguistics and algorithms to transform or enrich the text in such a way that it delivers greater value.