Image Quality in Digital Pathology

55
Image Quality in Digital Pathology (from a pathologist’s perspective) Jonhan Ho, MD, MS

description

Image Quality in Digital Pathology. (from a pathologist’s perspective) Jonhan Ho, MD, MS. Disclosure. Image Quality: define/measure. Image quality is good enough if:. It has a resolution of 0.12345 μ /pixel It is captured in XYZ color space/pixel depth - PowerPoint PPT Presentation

Transcript of Image Quality in Digital Pathology

Page 1: Image Quality in Digital Pathology

Image Quality in Digital Pathology

(from a pathologist’s perspective)Jonhan Ho, MD, MS

Page 2: Image Quality in Digital Pathology

Disclosure

Page 3: Image Quality in Digital Pathology
Page 4: Image Quality in Digital Pathology
Page 5: Image Quality in Digital Pathology
Page 6: Image Quality in Digital Pathology

Image Quality: define/measure

Page 7: Image Quality in Digital Pathology

Image quality is good enough if:

• It has a resolution of 0.12345 μ/pixel• It is captured in XYZ color space/pixel depth• It has a MTF curve that looks perfect• It has a focus quality score of 123• Has a high/wide dynamic range

Page 8: Image Quality in Digital Pathology

What is “resolution”?

• Spatial resolution• Sampling period• Optical resolution• Sensor resolution• Monitor resolution• New Year’s resolution???????

Page 9: Image Quality in Digital Pathology

Optical resolution

• Theoretical maximum resolution of a 0.75 NA lens is 0.41μ. 1.30 NA – 0.23μ.

• Has NOTHING to do with magnification! (we will get to that later.)

NA

R61.0

Page 10: Image Quality in Digital Pathology

Depth of Field

• As aperture widens– Resolution improves– Depth of field narrows• Less tissue will be in focus

Page 11: Image Quality in Digital Pathology

Image quality is good enough if:

• It has a resolution of 0.12345 μ/pixel• It is captured in XYZ color space/pixel depth• It has a MTF curve that looks perfect• It has a focus quality score of 123• Has a high/wide dynamic range

Page 12: Image Quality in Digital Pathology

Image quality is good enough if it is:

• “Sharp”• “Clear”• “Crisp”• “True”• “Easy on the eyes”

Page 13: Image Quality in Digital Pathology

Image quality is good enough if it is:

• “Sharp”• “Clear”• “True”

Page 14: Image Quality in Digital Pathology

Image quality is good enough if:

• You can see everything you can see on a glass slide

Page 15: Image Quality in Digital Pathology
Page 16: Image Quality in Digital Pathology
Page 17: Image Quality in Digital Pathology

Image quality is good enough if:

• I can make a diagnosis from it

Page 18: Image Quality in Digital Pathology
Page 19: Image Quality in Digital Pathology
Page 20: Image Quality in Digital Pathology

Image quality is good enough if:

• I can make as good a diagnosis from it as I can glass slides.– This is a concordance study

• OK, but how do you measure this?!?!?!?!?!

Page 21: Image Quality in Digital Pathology

Gold standard = Another Diagnosis

Glass

Observer

Original?

Page 22: Image Quality in Digital Pathology

Concordance validation

• Some intra-observer variability• Even more interobserver variability• Order effect• “great case” effect

Page 23: Image Quality in Digital Pathology

Concordance validation

• Case selection– Random, from all benches?– Enriched, with difficult cases?– Presented with only initial H&E?• Allow ordering of levels, IHC, special stains?• If so, how can you compare with the original diagnosis?

– Presented with all previously ordered stains?• If so, what about diagnosis bias?

– How old of a case to allow?

Page 24: Image Quality in Digital Pathology

Concordance validation

• Subject selection– Subspecialists? Generalists?– Do all observers read all cases, even if they are not

accustomed to reading those types of cases?– Multi-institutional study• Do observers read cases from other institutions?• Staining/cutting protocol bias

Page 25: Image Quality in Digital Pathology

Concordance validation

• Measuring concordance– Force pathologist to report in discrete data

elements?• This is not natural! (especially in inflammatory

processes!)• What happens if 1 data element is minimally

discordant?

– Allow pathologist to report as they normally do?• Free text – who decides if they are concordant? How

much discordance to allow? What are the criteria?

Page 26: Image Quality in Digital Pathology

Concordance study bottom line

• Very difficult to do with lots of noise• Will probably conclude that can make equivalent

diagnoses• At the end, we will have identified cases that are

discordant, but what does that mean?– What caused the discordances?

• Bad images? If so what made them bad?• Familiarity with digital?• Lack of coffee?!?!?!

• Still doesn’t feel like we’ve done our due diligence – what exactly are the differences between glass and digital?

Page 27: Image Quality in Digital Pathology

PERCEPTION = REALITY

Page 28: Image Quality in Digital Pathology

PERCEPTION = QUALITY

“Sharp, clear, true”

Page 29: Image Quality in Digital Pathology

Psychophysics

• The study of the relationship between the physical attributes of the stimulus and the psychological response of the observer

Page 30: Image Quality in Digital Pathology

What we need is -

Image

Image quality

Observer Performance

Page 31: Image Quality in Digital Pathology

Images, image quality and observer performance: new horizons in radiology lecture. Kundel HL. Radiology. 1979 Aug;132(2):265-71

Page 32: Image Quality in Digital Pathology

Kundel on image quality

• “The highest quality image is one that enables the observer to most accurately report diagnostically relevant structures and features.”

Page 33: Image Quality in Digital Pathology
Page 34: Image Quality in Digital Pathology

Receiver Operator Curve (ROC)

Page 35: Image Quality in Digital Pathology

Conspicuity index formula

• K = f(Size, contrast, Edge Gradient/surround complexity)

• Probability of detection = f(K)

Page 36: Image Quality in Digital Pathology

Kundel, 1979

• “Just as a limited alphabet generates an astonishing variety of words, an equally limited number of features may generate an equally astonishing number of pictures.”

Page 37: Image Quality in Digital Pathology

Can this apply to pathology?

• What is our alphabet? MORPHOLOGY!– Red blood cells– Identify inflammation by features

• Eosinophils• Plasma cells

– Hyperchromasia, pleomorphism, NC ratio– Build features into microstructures and macrostructures– Put features and structures into clinical context and

compare to normal context– Formulate an opinion

Page 38: Image Quality in Digital Pathology
Page 39: Image Quality in Digital Pathology
Page 40: Image Quality in Digital Pathology
Page 41: Image Quality in Digital Pathology
Page 42: Image Quality in Digital Pathology

Advantages of feature based evaluation

• Better alleviates experience bias, context bias• Can better perform interobserver concordancy• Connects pathologist based tasks with

measurable output understandable by engineers

• Precedent in image interpretability (NIIRS)

Page 43: Image Quality in Digital Pathology

NIIRS 1

“Distinguish between major land use classes (agricultural, commercial, residential)”

Page 44: Image Quality in Digital Pathology

NIIRS 5

“Identify Christmas tree plantations”

Page 45: Image Quality in Digital Pathology

Disadvantages of feature based evaluation

• Doesn’t eliminate the “representative ROI” problem

• Still a difficult study to do– How to select features? How many?– How to determine gold standard?– What about features that are difficult to discretely

characterize? (“hyperchromasia”, “pleomorphism”)

Page 46: Image Quality in Digital Pathology

Bottom line for validation

• All of these methods must be explored as they each have their advantages and disadvantages– Technical– Diagnostic concordance– Feature vocabulary comparison

Page 47: Image Quality in Digital Pathology

Image perception - Magnification

• Ratio• Microscope– Lens– Oculars

• Scanner– Lens– Sensor resolution– Monitor resolution– Monitor distance

Page 48: Image Quality in Digital Pathology

40X magnification from object to

sensor

1 pixel = 10 µm at the sensor1 pixel = 0.25 µm at the sample

10/0.25 = 40X

270 µm pixel pitch of monitor

~27X magnification from sensor to

monitor

1 pixel =270 µm at the monitor 1 pixel = 10 µm at the sensor

270 / 10 = ~27X

= 1080X TOTAL magnification from object to monitor

This is the equivalent of a 108X objective on a microscope!!??

Magnification at the monitor

Page 49: Image Quality in Digital Pathology

Scan Type

Magnification Effective Viewing Magnification (at 1:1)

Manual Scope Equivalent Objective Magnification

Object to

Sensor

Sensor to

MonitorTOTAL 10” 24” 48” 10” 24” 48”

20X 20 27 540 540 225 112.5 ~54x ~22.5x ~11.3x

40X 40 27 1080 1080 450 225 ~108x ~45x ~22.5x

Near point = 10”

What if the sensor was obscenely high resolution?

Page 50: Image Quality in Digital Pathology

Other things that cause bad images

• Tissue detection• Focus

Page 51: Image Quality in Digital Pathology

Tissue detection

Page 52: Image Quality in Digital Pathology

What about Phantoms?

Page 53: Image Quality in Digital Pathology

One final exercise in image perception

Page 54: Image Quality in Digital Pathology