Quality Assurance

FAQ: How do you assure the quality of your categorization results?

The quality of every categorisation is measured by every single category as some categories are more important than others (e.g. salary detection is much more difficult but also much more important for our customers than payments in supermarkets).

For each category we measure the classical precision and recall KPIs and compute the F1 score which is a balanced average of recall and precision. “Recall” - roughly speaking - measures the capability of the machine to detect the desired category (“how many of the expected category occurrences are detected”), while “precision” measures the accuracy of the detected occurrences.

For instance, typical F1 scores in January 2020 were:

Category

Description

F1 Score

A.5.4

Payments to collection offices

98.7%

E.1.1

Salary

98.9%

K.2.1

Rent

92.0%

Fintecsystems has implemented a quality assurance process comprising the following 9 steps.

1 Immediate Support

Every customer can send direct feedback to the categorization team which takes care of the request. Most requests are of explanatory nature, some are problem reports that the team directly takes care of.

2 Random sample testing

Our categorization team does random sampling checks of categorized turnovers and sets the desired outcome. Several employees have to come to the same conclusion to settle a manual categorization decision (majority decision). This process looks at thousands of account turnovers every month and the results are processed in a 30-days-round-robin process.

3 Computing statistics

The manual majority decisions are used to compute recall, precision and F1 score for every category. This helps us to measure the quality of our engine and to steer our work.

4 Four-eyes-principle

As in our software development department, changes of the categorization logic cannot be done by only one person. A second expert has to check the changes in the categorization engine and accept them as useful and valid.

5 Automated tests

Every change in the categorization engine has to pass an automated test. If this automated test fails, the change cannot go productive.

6 Before-after-comparison

Every change in the categorization engine has to be checked by a team member in a before-after-comparison manner, meaning that the “before” engine and “after” engine are run on a big sample of real account turnovers and changes are displayed to be checked by the expert.

7 Versioning

Every version of the categorization engine is strictly versioned like any other piece of software that we are producing. If there should be a problem in a new version, we can return to a previous version any time and within seconds.

8 Problem case discussion

The categorization team does weekly or bi-weekly phone meetups to discuss the handling of edge and problem cases. Every edge or problem case is documented.

9 Team agreement measurement

All manual categorization decisions of all team members are monitored in a way that we can detect deviations from the “decision norm”. Comparing a vast number of manual decisions done by several team members shows individual misunderstanding or training needs.

Last updated