2025-09-09.

My experience: fine-tuning for info extraction

During Q1 and Q2 of 2025, I was cooking a model for the task of Industrial Product Property Extraction given product description (textual) - with the same training data as BERT-CRF, the fine-tuned QWen-2.5-7B models failed to surpass BERT-CRF on supply-side product description (which is longer and more complex) and succeeded on surpass on demand-side product description (which is shorter and more succinct) a little bit.

The fine-tuning recipe is simple as follows.

- Method: Supervised Fine-Tuning (SFT)
- Data format:
  - Input: {Instructions (Task desc + Multi req + Output format)}
  - Output: {JSON properties as PropName-PropValList key-value pairs}

Phenomenon 1: Overfitting to position bias

xxx