Questions tagged [natural-language-processing]
For questions related to natural language processing (NLP), which is concerned with the interactions between computers and human (or natural) languages, in particular how to create programs that process and analyze large amounts of natural language data.
22 questions from the last 365 days
1
vote
0
answers
12
views
Why do transformer models sometimes produce fluent but logically inconsistent answers even when retrieval provides the correct context?
I understand that transformer-based language models can generate highly fluent responses, and that retrieval-augmented generation (RAG) is often used to improve factual grounding by supplying relevant ...
0
votes
1
answer
26
views
Are there AI architectures where one model generates reasoning and another model verifies or monitors the reasoning process?
I am a student who has recently started learning about artificial intelligence and reasoning systems, so I apologize in advance if this question is already well known in the literature.
Many modern ...
1
vote
2
answers
86
views
Why do large language models hallucinate facts even when trained on large datasets?
Large language models such as GPT, LLaMA, and Claude are trained on massive datasets and can generate highly coherent text. However, they still frequently produce incorrect or fabricated information, ...
0
votes
1
answer
13
views
How to design similarity search for mechanical products
I have a website which shows different products about machines and its different parts.
There are 10000 of product pages, and want to build a functionality which shows similar product pages and ...
2
votes
1
answer
58
views
System Prompts & LLM [closed]
I’m currently building an AI chatbot that converts natural language user queries into accurate SQL queries, executes them on a database, and returns the results in a simple, readable format.
At the ...
4
votes
1
answer
146
views
What are the current state-of-the-art techniques for reducing hallucinations in large language models?
I’m studying how modern large language models (LLMs) generate factual and verifiable outputs. Despite improvements in training data quality and model alignment, hallucinations still occur.
My question ...
3
votes
4
answers
3k
views
Can AI hallucination be regarded as a machine error?
Can AI hallucination be regarded as an error in which the model is incapable of sacrificing the quality of language of its response for its integrity, due to its discrete, all or none nature?
An ...
3
votes
1
answer
54
views
BERT [CLS] token captures context by attending to all other tokens -- isn't this true for vocab tokens too?
I've frequently seen it mentioned that the embedding of the [CLS] token from BERT can capture the context of the sequence because it "attends to all the other tokens". But BERT implements ...
2
votes
2
answers
212
views
If LLMs like OpenAI / DeepSeek / Gemini exist, why do we still need ML or NLP libraries, now and in the future?
I’m new to AI and NLP, and I’m trying to understand how different tools fit together.
Large Language Models (LLMs) like OpenAI, DeepSeek, or Gemini can already handle many NLP tasks text ...
0
votes
1
answer
97
views
How do I manage a discussion between two characters in code?
I want to code (C++) a method allowing a character C1 to ask or request something from another character C2.
The answer of C2 will be environment related:
does it knows the thing C1 is looking for?
...
0
votes
0
answers
49
views
Extracting services mentioned in short reports — rules vs ML?
I’m trying to identify which home services are present vs explicitly excluded in short, free-text reports. I also need to normalize synonyms (e.g., “pressure washing” → “power washing”).
Goal: decide ...
2
votes
1
answer
107
views
Are traditional NLP tasks solve well with modern transformer/llm technologies?
Before the LLM explosion, the tradidional NLP tasks, such as parsing, coreference resolution, translation to logical representation, temporal and event sequence resolution -- all have approached a ...
0
votes
0
answers
71
views
How can I give context to the BLIP model when generating captions?
I'm using HuggingFace's 'blip-image-captioning-base' model for image captioning. I trained it on both existing and domain-specific datasets I created specifically for generating Turkish language ...
2
votes
0
answers
53
views
How can I integrate BERT Tokenizer into BLIP model for image captioning?
Lately, I've been working on generating alt text for images using the BLIP model. The model I use is "blip-image-captioning-base" from HuggingFace. However, to generate alt text in Turkish ...
0
votes
0
answers
52
views
How can a symbolic "traveler equation" help AI detect signal-carrying works across time?
**In my personal project, "The Circular Vision: an equation to find signal‑carrying humans", I propose a symbolic framework to think about how some human works (songs, poems, scientific ...