Skip to main content
. 2025 Mar 27;27:e65537. doi: 10.2196/65537

Table 3.

Zero-shot and few-shot learning performance comparison of GPT-4 versus BERT and Lattice-LSTM models on Chinese medical entity extraction dataset.

Model CMeEEa

Precision Recall F1-score
BERTb-base 63.08 64.08 62.11
BERT-wwm 61.5 61.29 61.72
Lattice-LSTMc 46.34 43.60 49.44
Med-BERT 53.33 47.58 60.66
GPT-4 and Zero shot 64.07 68.97 59.82
GPT-4 and Few shot 65.31 64.89 65.73

aCMeEE: Chinese Medical Entity Extraction.

bBERT: Bidirectional Encoder Representations from Transformers.

cLSTM: long short-term memory.