From Naive Bayes to Large Language Models: A New Era of Text-Based Classification

Rajat Verma Co-Author
 
Abdul Wasay First Author
Autodesk
 
Abdul Wasay Presenting Author
Autodesk
 
Monday, Aug 4: 11:30 AM - 11:35 AM
2166 
Contributed Speed 
Music City Center 
Text classification has been central to data science, with models like Naive Bayes serving as foundational tools. While effective for short, structured data, traditional classifiers struggled with long-form text due to methods like TF-IDF, which failed to capture context and long-term dependencies. Tasks like sentiment analysis of reports or categorizing legal documents often led to inefficiencies. Large language models (LLMs) have transformed this landscape. Transformer architectures excel at understanding context within text, enabling accurate classification for complex tasks such as identifying topics in policy documents or classifying unstructured medical records. Further, smaller, fine-tuned LLMs like BERT provide scalable, cost-effective solutions without sacrificing performance compared to larger general-purpose LLMs like GPT-4o which involve higher costs for both training and inference. This talk focuses on how such smaller LLMs can be utilized to solve classification problems at a low cost with high accuracy. It'll highlight strategies for deploying such models efficiently so that they empower data scientists and statisticians to tackle challenges in modern text analytics.

Keywords

Text classification

Large language models (LLMs)

Transformer architectures

BERT

Cost-effective solutions


Modern text analytics 

Main Sponsor

Section on Statistics and Data Science Education