To classify exudates (spots) representing diabetic retinopathy (DR) in fundus images, I've taken the following two approaches:

1) The sum is taken of all objects for each features.  These sums are used as image level features.  This classifier would then produce a single label for an unknown image, either pathological or healthy.

2) Every object in each image has features extracted that are used for training.  The label assigned to each object feature set is the DR grade for that image.  So, if an image is pathological (1), then all features extracted from that image are assigned label 1, and conversely for healthy images.  This classifier would then produce a set of labels for an unknown object, classifying each object contained in the unknown image.  We could then say if any object in the image is pathological, then the image is pathological.

I'm receiving better results for method 2 than for method 1.  Is this an acceptable measure of classification performance to be used in publishable research?

More John Jenkinson's questions See All
Similar questions and discussions