Title:

Learn-to-Retrieve-and-Generate for Article Generation

Poster

Preview Converted Images may contain errors

Abstract

Advances in Natural Language Processing (NLP) and Information Retrieval (IR) models over the past several years have given rise to improvements in text retrieval and generation tasks. We look to leverage one such model, Google's Bidirectional Encoder Representation from Transformers (BERT) for the task of Wikipedia article generation via next-paragraph-prediction. Using a pre-trained implementation of BERT, we fine-tune on top of this model to sequentially predict and retrieve paragraphs from an article's set of paragraphs. Articles are generated continuously, basing the prediction for the next paragraph on the generated article thus far, rather than on the previously retrieved paragraph. This work looks to serve as a basis for further paragraph-level retrieval techniques, as well as text-level article generation.

Authors

First Name Last Name
Joseph Guidoboni

File Count: 1


Leave a comment

Comments are viewable only by submitter



Submission Details

Conference URC
Event Interdisciplinary Science and Engineering (ISE)
Department Computer Science (ISE)
Group Data Science
Added April 25, 2021, 10:12 p.m.
Updated April 19, 2023, 10:39 a.m.
See More Department Presentations Here