Has LLM killed traditional NLP?
Large Language Models (LLMs) streamline Natural Language Processing by using zero-shot prompts, reducing the need for extensive training data and retraining, potentially challenging traditional NLP methods' relevance and efficiency.
Read original articleThe article discusses the impact of Large Language Models (LLMs) on traditional Natural Language Processing (NLP) methods. Traditionally, NLP has involved breaking down tasks into smaller problems, such as text classification and Named Entity Recognition (NER), requiring extensive training data and model retraining for new intents. For instance, to classify intents like check-in time inquiries, developers must create detailed examples and retrain models as new intents are added. However, with the advent of LLMs like ChatGPT, the process has become more streamlined. LLMs can handle various NLP tasks using zero-shot prompts, allowing users to input questions and intents without needing extensive examples or retraining. This shift raises questions about the future of traditional NLP techniques, as LLMs offer a more efficient and flexible approach to solving language-related challenges.
- LLMs simplify NLP tasks by using zero-shot prompts, reducing the need for extensive training data.
- Traditional NLP methods require detailed examples and retraining for new intents, making them more time-consuming.
- The rise of LLMs may challenge the relevance of traditional NLP techniques in the industry.
- LLMs can potentially handle a wider range of language tasks more efficiently than traditional models.
Related
One prompt? Fair. 10? Still ok. 100? You're pushing it. 10M - get help.
I have worked on AWS Connect (online call center) and Amazon Lex (the backing NLP engine) projects.
Before LLMs, it was a tedious process of trying to figure out all of the different “utterances” that people could say and the various languages you had to support. With LLMs, it’s just prompting
https://chatgpt.com/share/678bab08-f3a0-8010-82e0-32cff9c0b4...
I used something like this using Amazon Bedrock and a Lambda hook for Amazon Lex. Of course it wasn’t booking a flight. It was another system
The above is a simplified version. In the real world , I gave it a list of intents (book flights, reserve a room, rent a car) and properties - “slots” - I needed for each intent.
Text classification, clustering, named entity recognition, etc. are NLP tasks. LLMs can perform these tasks. ML models that are not LLMs (or even not deep learning models) can also perform these tasks. Is the author perhaps asking if the concept of a "completion" has replaced all of these tasks?
When I hear "traditional NLP" I think not of the above types of tasks but rather the methodology employed for performing them. For example, building a pipeline to do stemming/lemmatization, part of speech tagging, coreference resolution, etc. before the text gets fed to a classifier model. This was SOTA 10 years ago but I don't think many people are still doing it today.
''E's got a 'ittle box 'n a big 'un,' she said, 'wit' th' 'ittle 'un 'bout 2'×6". An' no, y'ain't cryin' on th' "soap box" to me no mo, y'hear. 'Cause it 'tweren't ever a spec o' fun!' I says to my frien'.
The library is integrated into my Markdown editor, KeenWrite (https://keenwrite.com/), to correctly curl quotation marks into entities before passing them over to ConTeXt for typesetting. While there are other ways to indicate opening and closing quotation marks, none are as natural to type in plain text as straight quotes. I would not trust an LLM curl quotation marks accurately.For the curious, you can try it at:
https://whitemagicsoftware.com/keenquotes/
If you find any edge cases that don't work, do let me know. The library correctly curls my entire novel. There are a few edge cases that are completely ambiguous, however, that require semantic knowledge (part-of-speech tagging), which I haven't added. PoS tagging would be a heavy operation that could prevent real-time quote curling for little practical gain.
The lexer, parser, and test cases are all open source.
https://gitlab.com/DaveJarvis/KeenQuotes/-/tree/main/src/mai...
https://www.microsoft.com/en-us/research/blog/lazygraphrag-s...
https://freedium.cfd/https://medium.com/altitudehq/is-tradit...
An equivalently funny attitude seems to be the "natural language will replace programming languages". Let's see how that one will work out when the hype is over.
Every business is kind of a unicorn in its problems NLP is a small part of it. Like even if it did perform cheaply enough to do NLP, how would you replace parts like: 1. Evaluation system that uses Calibration(Human labels) 2. Ground Truth Collection(Human + sometimes semi automated) 3. QA testing by end users.
Even if LLMs made it easier to do NLP there are correlations with the above which means your NLP process is hugely influenced so much that you still need an engineer. If you have an engineer who only for doing NLP and nothing else you are quite hyper specialized like to the extent you are only building planes 0.01%: of the engineering work out there.
FYI - If anyone doesn't know the difference between the two or has no idea what NLP or an LLM is, this has a good breakdown: https://medium.com/@melindaboone80722/nlp-vs-llm-b339abdc651...
We've completely replaced that with LLMs. We still use our own DNNs for certain tasks, but not for NLP.
- Depending on the complexity of the task and the required results, SVMs or BERT can be enough in many cases and take much lower resources, especially if there is a lot of training data available. Training these models with LLM outputs could also be an interesting approach to achieve this.
- When resources are constrained or latency is important.
- In some cases, there may be labeled data in certain classes that have no semantic connection between them, e.g. explaining the class to LLMs could be tricky.