Title:

BLENS: Biomedical Literature Extraction & Scoring System

Poster

Preview Converted Images may contain errors

Abstract

Systematic reviews require comprehensive retrieval across multiple databases, yet no publicly available tool exists for automated cross-database literature quality scoring. BLENS retrieves records from PubMed, ClinicalTrials.gov, medRxiv, and bioRxiv, scoring all output via a ML pipeline validated on Third Molar Extraction (TME) literature and keywords. Text is encoded into dense vectors using Sentence-BERT and scored by cosine similarity to a domain-specific query. A semi-supervised anchoring layer uses 5 expert-validated gold-standard studies as labeled seeds to recalibrate scores across all records. A rubric then scores against 5 weighted inclusion criteria gathered from a published TME meta-analysis, combining these scores into three bins tiered High, Medium, or Low. Applied to 1,795 records, Blens achieves 100% recall on gold-standard studies, with composite scores improving from 0.59-0.72 at baseline to 0.80-0.99 after anchoring. HDBSCAN clustering independently confirmed results, separating preprints into off-topic clusters due to a vague retrieval from a weak API. Anomaly rates of 25% (bioRxiv) and 15.6% (medRxiv) versus 0.8% (PubMed) and 0% (ClinicalTrials) reflects the poor retrieval of preprints due to API query limits. BLENS is a proof of concept, where precision measurement via manual audit of top-rank records is the next critical step.

Authors

First Name Last Name
Tyler Simone

Advisors:

Full Name
Marek Petrik
Matthew Magnusson

File Count: 1


Leave a comment

Comments are viewable only by submitter



Submission Details

Conference URC
Event Interdisciplinary Science and Engineering (ISE)
Department Computer Science (ISE)
Group Computer Science - Independent Projects
Added April 20, 2026, 7:25 p.m.
Updated April 20, 2026, 7:25 p.m.
See More Department Presentations Here