Transformer Neural Networks for Natural Language Processing

Loading...
Thumbnail Image

Date

Authors

Tembo, Malinga

Mark

B

Journal Title

Journal ISSN

Volume Title

Publisher

Vysoké učení technické v Brně. Fakulta elektrotechniky a komunikačních technologií

ORCID

Abstract

This thesis outlines a retrieval-augmented generation (RAG) pipeline tailored specifically for domain-specific documentation, with particular focus on telecoms config guides. The pipeline outlined here integrates structural pre-processing, dense semantic indexing, and controlled prompt composition to improve factual coherence in LLM responses. The primary contribution is the design of a preprocessing method, based on which semi-structured guides are converted into hierarchically divided JSON files, to enable context-aware retrieval over a vector database. The system is evaluated on 45 queries comprising question answering, summarization, and CLI code generation, on the Arista EOS 4.17.0F guide used as a test corpus. Empirical performance shows retrieval drastically reduces hallucination and increases semantic relevance, with GPT-4 + RAG producing 0.82 accuracy and 0.02 hallucination rate. For contrast, TinyLlama returned nonsensical output for over one-third of inputs. These findings demonstrate the effectiveness of grounded retrieval in aligning model output with technical reference material and illustrate the benefit of structured input preparation for high-fidelity generation in documentation-heavy applications.
This thesis outlines a retrieval-augmented generation (RAG) pipeline tailored specifically for domain-specific documentation, with particular focus on telecoms config guides. The pipeline outlined here integrates structural pre-processing, dense semantic indexing, and controlled prompt composition to improve factual coherence in LLM responses. The primary contribution is the design of a preprocessing method, based on which semi-structured guides are converted into hierarchically divided JSON files, to enable context-aware retrieval over a vector database. The system is evaluated on 45 queries comprising question answering, summarization, and CLI code generation, on the Arista EOS 4.17.0F guide used as a test corpus. Empirical performance shows retrieval drastically reduces hallucination and increases semantic relevance, with GPT-4 + RAG producing 0.82 accuracy and 0.02 hallucination rate. For contrast, TinyLlama returned nonsensical output for over one-third of inputs. These findings demonstrate the effectiveness of grounded retrieval in aligning model output with technical reference material and illustrate the benefit of structured input preparation for high-fidelity generation in documentation-heavy applications.

Description

Citation

TEMBO, M. Transformer Neural Networks for Natural Language Processing [online]. Brno: Vysoké učení technické v Brně. Fakulta elektrotechniky a komunikačních technologií. 2025.

Document type

Document version

Date of access to the full text

Language of document

en

Study field

bez specializace

Comittee

doc. Ing. Jan Jeřábek, Ph.D. (místopředseda) M.Sc. Sara Ricci, Ph.D. (člen) Ing. Martin Štůsek, Ph.D. (člen) Ing. Pavel Paluřík (člen) Ing. Willi Lazarov (člen) prof. Ing. Miroslav Vozňák, Ph.D. (předseda)

Date of acceptance

2025-06-09

Defence

Student presented the results of his thesis and the committee got familiar with reviewer's report. Student defended his Diploma Thesis and answered the questions from the members of the committee and the reviewer.

Result of defence

práce byla úspěšně obhájena

DOI

Collections

Endorsement

Review

Supplemented By

Referenced By

Citace PRO