Generative AI and unstructured audio data for precision public health.
Anibal J., Landa A., Nguyen H., Daoud V., Le T., Huth H., Song M., Peltekian A., Shin A., Hazen L., Christou A., Rivera J., Morhard R., Brenner J., Bagci U., Li M., Bensoussan Y., Clifton D., Wood B.
In this study, transcribed videos about personal experiences with COVID-19 were used for variant classification. The o1 LLM was used to summarize the transcripts, excluding references to dates, vaccinations, testing methods, and other variables that were correlated with specific variants but unrelated to changes in the disease. This step was necessary to effectively simulate model deployment in the early days of a pandemic when subtle changes in symptomatology may be the only viable biomarkers of disease mutations. The embedded summaries were used for training a neural network to predict the variant status of the speaker as "Omicron" or "Pre-Omicron", resulting in an AUROC score of 0.823. This was compared to a neural network model trained on binary symptom data, which obtained a lower AUROC score of 0.769. Results of the study illustrated the future value of LLMs and audio data in the design of pandemic management tools for health systems.