Skip to main content
OpenConf small logo

Providing all your submission and review needs
Abstract and paper submission, peer-review, discussion, shepherding, program, proceedings, and much more

Worldwide & Multilingual
OpenConf has powered thousands of events and journals in over 100 countries and more than a dozen languages.

Traditional Machine Learning Models Versus Small Language Models For Classification of Orthoses, Prostheses, and Special Materials In Invoices

This paper investigates the use of traditional machine learning models versus Small Language Models (SLMs) for classifying Orthoses, Prostheses, and Special Materials (OPMEs) in invoices, a critical problem for healthcare auditing. Building on the OPME dictionary used by the OPMinEr system, a controlled experiment is conducted comparing a Linear SVM classifier with TF–IDF representations against a fine-tuned BERTimbau model, in both a binary task (OPME vs non-OPME) and a multiclass task (orthosis, prosthesis, special material, others). The results show that the SVM achieves accuracy and macro F1 scores around 0.98–0.99, slightly outperforming BERTimbau in both tasks, while requiring orders of magnitude less training and inference time. These findings suggest that, in the specific context of healthcare invoices, well-tuned traditional models remain the most effective and cost-efficient choice for large-scale deployment, whereas SLMs are most promising as complementary tools in research settings and for potential transfer to related classification tasks.

Natan Matos
Universidade Federal de Sergipe
Brazil

Methanias Colaço
Universidade Federal de Sergipe
Brazil