Inproceedings3985: Unterschied zwischen den Versionen
Nu4126 (Diskussion | Beiträge) (Die Seite wurde neu angelegt: „{{Publikation Erster Autor |ErsterAutorNachname=Yuan |ErsterAutorVorname=Shuzhou }} {{Publikation Author |Rank=2 |Author=Michael Färber }} {{Inproceedings |Re…“) |
Nu4126 (Diskussion | Beiträge) |
||
Zeile 19: | Zeile 19: | ||
{{Publikation Details | {{Publikation Details | ||
|Abstract=Large language models (LLMs) have been widely employed for graph-to-text generation tasks. However, the process of finetuning LLMs requires significant training resources and annotation work. In this paper, we explore the capability of generative models to generate descriptive text from graph data in a zero-shot setting. Specifically, we evaluate GPT-3 and ChatGPT on two graph-to-text datasets and compare their performance with that of finetuned LLM models such as T5 and BART. Our results demonstrate that generative models are capable of generating fluent and coherent text, achieving BLEU scores of 10.57 and 11.08 for the AGENDA and WebNLG datasets, respectively. However, our error analysis reveals that generative models still struggle with understanding the semantic relations between entities, and they also tend to generate text with hallucinations or irrelevant information. As a part of error analysis, we utilize BERT to detect machine-generated text and achieve high macro-F1 scores. We have made the text generated by generative models publicly available. | |Abstract=Large language models (LLMs) have been widely employed for graph-to-text generation tasks. However, the process of finetuning LLMs requires significant training resources and annotation work. In this paper, we explore the capability of generative models to generate descriptive text from graph data in a zero-shot setting. Specifically, we evaluate GPT-3 and ChatGPT on two graph-to-text datasets and compare their performance with that of finetuned LLM models such as T5 and BART. Our results demonstrate that generative models are capable of generating fluent and coherent text, achieving BLEU scores of 10.57 and 11.08 for the AGENDA and WebNLG datasets, respectively. However, our error analysis reveals that generative models still struggle with understanding the semantic relations between entities, and they also tend to generate text with hallucinations or irrelevant information. As a part of error analysis, we utilize BERT to detect machine-generated text and achieve high macro-F1 scores. We have made the text generated by generative models publicly available. | ||
+ | |Download=RANLP_2023_short_r2.pdf | ||
|Forschungsgruppe=Web Science | |Forschungsgruppe=Web Science | ||
}} | }} |
Version vom 27. Juli 2023, 11:35 Uhr
Evaluating Generative Models for Graph-to-Text Generation
Evaluating Generative Models for Graph-to-Text Generation
Published: 2023
September
Buchtitel: Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP 2023)
Verlag: ACL Anthology
Erscheinungsort: Varna, Bulgaria
Organisation: RANLP 2023
Referierte Veröffentlichung
BibTeX
Kurzfassung
Large language models (LLMs) have been widely employed for graph-to-text generation tasks. However, the process of finetuning LLMs requires significant training resources and annotation work. In this paper, we explore the capability of generative models to generate descriptive text from graph data in a zero-shot setting. Specifically, we evaluate GPT-3 and ChatGPT on two graph-to-text datasets and compare their performance with that of finetuned LLM models such as T5 and BART. Our results demonstrate that generative models are capable of generating fluent and coherent text, achieving BLEU scores of 10.57 and 11.08 for the AGENDA and WebNLG datasets, respectively. However, our error analysis reveals that generative models still struggle with understanding the semantic relations between entities, and they also tend to generate text with hallucinations or irrelevant information. As a part of error analysis, we utilize BERT to detect machine-generated text and achieve high macro-F1 scores. We have made the text generated by generative models publicly available.
Download: Media:RANLP_2023_short_r2.pdf
Natürliche Sprachverarbeitung, Künstliche Intelligenz