|J Pathol Inform 2013,
Inter-reader variability in follicular lymphoma grading: Conventional and digital reading
Gerard Lozanski1, Michael Pennell2, Arwa Shana'ah1, Weiqiang Zhao1, Amy Gewirtz1, Frederick Racke1, Eric Hsi3, Sabrina Simpson4, Claudio Mosse5, Shadia Alam6, Sharon Swierczynski7, Robert P Hasserjian8, Metin N Gurcan9
1 Department of Pathology, The Ohio State University, Columbus, OH, USA
2 Division of Biostatistics, College of Public Health, The Ohio State University, Columbus, OH, USA
3 Cleveland Clinic, Cleveland, OH, USA
4 Department of Pathology, Central Ohio Pathology Associates, Westerville, OH, USA
5 Vanderbilt University, Nashville TN, USA
6 Department of Pathology, Battle Creek, MI, USA
7 The Reading Hospital Medical Center, Reading PA, USA
8 Massachusetts General Hospital, Boston, MA, USA
9 Department of Biomedical Informatics, Ohio State University, Columbus, OH, USA
|Date of Submission||30-May-2013|
|Date of Acceptance||03-Sep-2013|
|Date of Web Publication||29-Oct-2013|
Metin N Gurcan
Department of Biomedical Informatics, Ohio State University, Columbus, OH
Source of Support: None, Conflict of Interest: None
| Abstract|| |
Context: Pathologists grade follicular lymphoma (FL) cases by selecting 10, random high power fields (HPFs), counting the number of centroblasts (CBs) in these HPFs under the microscope and then calculating the average CB count for the whole slide. Previous studies have demonstrated that there is high inter-reader variability among pathologists using this methodology in grading. Aims: The objective of this study was to explore if newly available digital reading technologies can reduce inter-reader variability. Settings and Design: In this study, we considered three different reading conditions (RCs) in grading FL: (1) Conventional (glass-slide based) to establish the baseline, (2) digital whole slide viewing, (3) digital whole slide viewing with selected HPFs. Six board-certified pathologists from five different institutions read 17 FL slides in these three different RCs. Results: Although there was relative poor consensus in conventional reading, with lack of consensus in 41.2% of cases, which was similar to previously reported studies; we found that digital reading with pre-selected fields improved the inter-reader agreement, with only 5.9% lacking consensus among pathologists. Conclusions: Digital whole slide RC resulted in the worst concordance among pathologists while digital whole slide reading selected HPFs improved the concordance. Further studies are underway to determine if this performance can be sustained with a larger dataset and our automated HPF and CB detection algorithms can be employed to further improve the concordance.
Keywords: Centroblast, follicular lymphoma, inter-reader variability, whole-slide images
|How to cite this article:|
Lozanski G, Pennell M, Shana'ah A, Zhao W, Gewirtz A, Racke F, Hsi E, Simpson S, Mosse C, Alam S, Swierczynski S, Hasserjian RP, Gurcan MN. Inter-reader variability in follicular lymphoma grading: Conventional and digital reading. J Pathol Inform 2013;4:30
|How to cite this URL:|
Lozanski G, Pennell M, Shana'ah A, Zhao W, Gewirtz A, Racke F, Hsi E, Simpson S, Mosse C, Alam S, Swierczynski S, Hasserjian RP, Gurcan MN. Inter-reader variability in follicular lymphoma grading: Conventional and digital reading. J Pathol Inform [serial online] 2013 [cited 2020 Jun 7];4:30. Available from: http://www.jpathinformatics.org/text.asp?2013/4/1/30/120747
| Introduction|| |
Follicular lymphoma (FL) is the second most common B-cell lymphoma affecting adults in the Western world.  FL is characterized by a highly variable clinical course that ranges from stable, indolent lymphoma that may subsequently progress to a more aggressive disease to a disease that behaves aggressively from the outset. Patients with indolent FL who are asymptomatic are usually not treated since there is no evidence that early therapy with currently available regimens provides benefit to these patients. ,,,,,, Such a "watch and wait" approach spares patients unnecessary therapy associated toxicity while allowing timely intervention when FL related symptoms develop and/or the disease progresses. ,, In contrast, patients who present with an aggressive form of FL at diagnosis often require immediate therapy to alleviate disease-related symptoms. ,, Understandably this marked clinical heterogeneity requires accurate risk stratification of all FL cases to guide the oncologist's clinical decision-making.
FL patients are risk-stratified according to clinical criteria using disease stage,  Follicular Lymphoma International Prognostic Index score  and histological grading.  Histological grading is performed according to the morphologic criteria of Mann-Berard, which have been adapted by the World Health Organization (WHO) classification.  In this grading system FL cases are divided into low grade (grade I and II) and high grade (grade IIIA and IIIB) based on the average count of centroblasts (CBs) per standard microscopic high power field (HPF). The CB count is manually performed by a pathologist in 10 random HPFs containing malignant follicles. FL cases with an average CB count from 0 to 15/HPF are classified as low grade and those with an average CB count of more than 15/HPF as grade III. Grade III is further subdivided into grade IIIA (demonstrating a mixed population of CBs and centrocytes) and grade IIIB (demonstrating a homogeneous population of CBs). As expected, this grading system performs well at the extreme ends of the spectrum with gradation of FL between grade I and grade IIIB being fairly reproducible. However, histological grading of FL cases at the interface between grade II and grade IIIA suffers from poor reproducibility even at the hands of expert hematopathologists.  This limitation of FL histological grading is very important since a large number of FL patients fall into a category bordering between low and high grade, can affect clinical management with a "watch and wait" approach versus chemotherapy.
Of the several factors impacting an accurate manual grading of FL based on CB count, the most important is the limitation of the human reader. Even when applying stringent criteria to categorize cells as CBs, human readers are prone to variable interpretation of specific cells as CBs and non-CBs that results in low accuracy and reproducibility of CB counts using unaided light microscope glass slide review. Moreover, since CB count is limited to 10 random HPF (by practical necessity) the heterogeneity of cell types present in a single FL can easily be under-represented. Recent development of high resolution imaging of histological slides and digital pathology techniques creates an opportunity to aid pathologists in accurate and reproducible FL grading. In this paper, we present the impact of digitization of FL cases on the accuracy and reproducibility of histological grading among six experienced hematopathologists. Similar to a previous study,  inter-pathologist variability in the glass slide readings was high as was the case when the pathologists viewed the whole slide digital images. However, superior inter-pathologist concordance was observed when pathologists were presented with the same HPFs and were obligated to mark cells counted as CBs.
Inter-reader variability in the grading of FL has previously been documented utilizing only conventional methods, i.e., glass slides, read under the microscope. In a study by The Non-Hodgkin's Lymphoma Classification Project, five pathologists reviewed 304 FL cases comprising grades I, II and III. On average, the individual pathologists agreed with the consensus diagnosis only 61-73% of the time (depending on grade) and immunophenotyping did not significantly add to the accuracy of the diagnosis.  In a similar study involving seven pathologists and 105 cases, Metter et al., found that for approximately half the cases (51%), the CB count range was more than 10 per HPF across pathologists and this range was more than 20/HPF for 29% of the cases.  With the recent widespread availability of digital whole-slide scanners, it is now possible to digitally capture, view, annotate and evaluate FL images. The use of digital images may help improve the accuracy and thus clinical utility of FL histologic grading.
| Subjects and Methods|| |
17 FL cases were selected from the archives of the first author's institution with IRB approval. These cases were randomly selected to represent different FL grades based on the existing pathology reports. All tissues were formalin-fixed, paraffin-embedded and hematoxylin and eosin (H and E) stained. One representative slide from each case was selected (by the first author) and used for this study, i.e., 17 slides were read. Each slide was scanned and converted to a digital image using an Aperio (Vista, CA) ScanScope scanner at ×40 magnification, which results in 0.23 μm per pixel resolution [Figure 1]. Following the acquisition of digital slides, one pathologist selected 10 HPFs (HPFs, approximately 0.159 mm 2 area) from each image. The HPFs were randomly selected from the areas representing malignant follicles in accordance with the WHO recommendations.
|Figure 1: Three different reading conditions (RCs): RC1 is conventional reading; in RC2, whole slide digital images are read by the pathologist; in RC3, selected high-power-fields are read by the pathologist|
Click here to view
Six board certified hematopathologists with at least 10 years of experience examined the 17 FL cases under three different reading conditions (RC1-3): Glass, digital whole slide and digital selected fields [Figure 1]. At least three months passed between reading experiments and prior to each reading the order of the slides was randomized to minimize the possibility of remembering cases.
RC1. Glass slide reading: This is the conventional and clinically accepted method of reading glass slides using a microscope following the standard WHO guidelines. The pathologists counted and recorded the number of CBs in 10 self-selected random fields representing malignant follicles according to the WHO recommendations and the project statistician computed the average number of CBs across the 10 fields. All the pathologists used the same type of microscope (Olympus Plan 40x-0263) equipped with a 40x dry objective (ocular: WH10x/22). The pathologists were instructed to use the WHO definition of CBs.  If more than 20 CBs were counted in a field, the count was rounded to 25 (if count between 21 and 30), 35 (if count between 31 and 40), 45 (if count between 41 and 50), or 55 (if count greater than 50) in computing the mean. Grade was determined using standard WHO guidelines: Average CBs per field ≤ 5 = Grade I; 6-15 = Grade II; >15 = Grade III. In order to make the counting practical, these limits were established; otherwise, pathologists cannot finish this study in a reasonable amount of time.
RC2. Digital whole slide reading: Digital whole slide readings followed a similar protocol to RC1 except that the readings took place on a computer rather than under a microscope using the ImageScope software [Figure 2]. Pathologists self-selected 10 HPFs and recorded the number of CBs for each selected field. The size of each selected area was adjusted to be equivalent to 0.159 mm 2 so that they were equivalent in the area to images viewed under the microscope although different in shape (circular under the microscope while rectangular on the computer screen). The equivalent area was calculated in pixels for digital reading. The workstation parameters were fixed and all the readers used the same software developed by our lab. In our experiments to standardize CB counting, we used one type of microscope and its digital equivalent for all readers and for all samples tested.
|Figure 2: Screen shot of the freely available commercial program (Imagescope, Aperio, Vista, CA) used for the digital evaluation slides for this study (reading condition 2 - RC2)|
Click here to view
RC3. Digital selected field reading: Finally, in the digital selected field readings, pathologists read the same fields randomly pre-selected by one of the pathologists. The selected fields were devoid of identifiers in order to blind the pathologists and the mean CBs per field was computed by the project statistician after data collection was completed. Selected images were marked using in-house developed software called CBMarker [Figure 3]. This software lets the pathologist connect to a secure server to mark individual CB locations by a simple mouse click on a selected HPF image. If a location is accidentally marked (i.e., wrong mouse click) then the erroneous marking can be easily deleted by clicking again on the same location. The image, marking location and marking pathologist information were recorded.
|Figure 3: Centroblast (CB) marker: The program to mark the locations of CBs on a high power field image reading condition 3|
Click here to view
Statistical Design and Methods
Variability in grade was determined using two metrics: (1) Number of cases for which the grade ranged from I to III across pathologists and (2) number of cases without a consensus (less than four pathologists agreed on grade I, II, or III). Exact Cochran's Q Tests were used to determine if either metric differed significantly across RCs and McNemar's tests were used to perform pairwise comparisons of the conditions.  In the pairwise comparisons, P values were corrected for multiple comparisons using Holm's method.  Kappa statistics were used to measure agreement between pathologists in WHO grade and clinically significant grade (Grade I or II vs. III). Landis and Koch guidelines were used to assess the level of agreement: <0 poor, 0-0.2 slight, 0.21-0.40 fair, 0.41-0.60 moderate, 0.61-0.80 substantial and > 0.80 almost perfect agreement.  We also calculated the number of cases for which each pathologist agreed with the consensus diagnosis of clinically significant grade (4 or more pathologists agreed on grade I/II or III) and compared results across RCs using repeated measures ANOVA.
In a separate set of analyses, we compared variability and performance in counting CBs across the three RC1-3. For each RC, the variability in the number of CBs per HPF was examined by calculating the range across pathologists. Pathologist performance in counting CBs was measured using the number of cases in which the pathologist's average CB count was more than 10 CBs greater than the mean across pathologists; a difference of 10 CBs is clinically significant as it could mean a two grade difference. The same approach to measuring variability and performance in counting CBs was used by Metter et al.  In both analyses, we compared the different RCs using repeated measures ANOVA and Tukey multiple comparisons of the means.  In the case of the range, the data were log transformed prior to analysis.
| Results|| |
[Table 1] summarizes the variability in WHO grade. When the pathologists had the freedom to select their own fields (glass and digital whole slide readings) over 35% of the cases had a grade range of I-III (i.e., at least one pathologist graded as I while at least one other pathologist graded as III) across pathologists and no consensus was reached for over 41%. However, when the pathologists were all enabled to read the same fields, there was only one case of non-consensus and two cases of grade range I-III, although only the first result was statistically significant (P < 0.01 for the difference across RCs).
|Table 1: Variability in WHO Grade (I, II, or III) across pathologists (6 pathologists, 17 cases) |
Click here to view
Inter-pathologist agreement in WHO grade was measured using pairwise Kappa statistics. As seen in [Table 2], agreement on grade I, II and III was best when the pathologists read in RC3 with a median Kappa of 0.64, which indicates substantial agreement and even the worst agreement in RC3 (0.41) was moderate according to the Landis and Koch guidelines.  In contrast, agreements in RC1 were mostly fair (0.21 ≤ Kappa ≤ 0.40) and slight (0 ≤ Kappa ≤ 0.2); and agreements in RC2 were mostly slight or poor (Kappa < 0). Furthermore, with two exceptions, the agreement between each pair of pathologists was greatest in RC3 (see [Figure 4] for RC1 vs. RC3 comparison; a similar trend was observed for RC2 vs. RC3). The average agreement in clinically significant grade (Grade I/II vs. III) was similar between RC1 and RC3 [Table 2] and neither was consistently superior to the other in terms of agreement of the individual pairs of pathologists [Figure 5].
|Figure 4: Graphical representation of difference in Kappa coefficients between reading condition (RC1) and RC3 readings: Agreement on grades I, II and III|
Click here to view
Performance of individual pathologists was measured in terms of agreement with consensus diagnosis of clinically significant grade. The consensus diagnoses for the RC1 and RC3 were identical: 14 grade I/II and 3 grade III. In the digital whole slide readings (RC2), the same 14 low grade cases were identified as low grade (i.e., as grades I or II), but no consensus was reached for the three cases identified as high grade in the other two RCs. The percentage of times each pathologist was in agreement with consensus is provided in [Table 3]. The average agreement with consensus was greatest for the selected field readings (RC3), but not significantly so (P = 0.331).
|Table 3: Number cases (%) in agreement with consensus diagnosis of clinically significant grade |
Click here to view
We also considered inter-pathologist variability and performance of pathologists in counting CBs. Histograms of the range in number of CBs per HPF by RC are provided in [Figure 6]. Ranges observed for the RC3 readings were smaller than both the RC1 and RC2 readings (P < 0.05). Pathologist performance in counting CBs was also best in the selected field readings. In the whole slide readings (RC1 and RC2), most pathologists were more than 10 CBs off from the overall mean for at least two cases [Table 4]. Under the selected field condition (RC3), only two pathologists provided counts that were more than 10 CBs from the overall mean, although the overall differences across RCs were only marginally significant (P = 0.09).
|Figure 6: Histograms of range in number of centroblasts/high power field across pathologists|
Click here to view
|Table 4: Number (%) of cases in which mean CB count was>10 cells different from the overall mean across pathologists |
Click here to view
| Discussion|| |
The most important finding of this study was that digital reading with pre-selected HPF improved -compared with the standard practice-the inter-reader agreement among pathologists grading FL cases and that whole slide digital reading worsened the consensus. In order to arrive at this conclusion, we designed an experiment with six board-certified pathologists from five different institutions and asked these pathologists to read 17 slides under three RCs. The first RC was the conventional reading, i.e., Pathologists read the slides according to the WHO criteria using their microscope. The second and the third RCs were digital whole slide readings without and with previously selected HPFs, respectively. While there was relatively poor consensus in conventional reading (lack of consensus in 41.2% of cases) similar to previously reported studies, we found that digital reading with pre-selected fields improved the inter-reader agreement, with only 5.9% lacking consensus among pathologists.
As explained in the Introduction and as the results of study again confirmed, current methods for grading FL suffer from high pathologist-to-pathologist variability. One of the major contributors to this variability is the fact that there are no specific guidelines for choosing the fields used to generate the CB count, which determines the grade. Hence, there is a great deal of heterogeneity in the location of the fields chosen. In this study, we have shown that the inter-subject variability in CB counts can be improved by enabling pathologists to view the same fields thereby improving agreement on grade. These results highlight the need for computer-aided diagnostic systems, which provide pathologists with consistent information obtained through objective algorithms, which may be used for the selection of fields or identification of cells or regions of interest.
There are active research programs in the computer-aided grading (CaG) of FL cases. [20 ],,,,,,,,,,,,,,,,,,,,,,,,,,, Particularly, there are efforts to examine the computational and human factor aspects of CaG, ,,,,,,, to develop multi-resolution and multi-classifier approaches to emulate expert cognitive functioning, ,,,, to investigate novel segmentation methods to identify follicles both in H and E and IHC images, ,,, methods to register multi-stain images  and detect cells. ,,,,, These studies showed that such systems could identify the most aggressive FL (grade III) with 98.9% sensitivity and 98.7% specificity and the overall classification accuracy of the system was 85.5%.  These methods were all designed to help pathologists perform the current grading system more accurately and consistently. While these efforts are on-going, this current study provided us with insight into the main factors that cause inter-reader variability and also what type of digital reading strategy should be followed.
Although digital slides are currently available and are widely used as teaching resources and for research purposes, they are not routinely used for clinical diagnosis. Current research is focusing on both how pathologists can use them in their clinical studies and what the optimal RCs should be. In this study, we used two digital RCs (RC2 - digital whole slide reading and RC3 - digital selected field reading). Our inter-pathologist agreement measures [Table 2] indicate that RC2 actually results in inferior results than current conventional reading. However, another digital reading strategy (RC3) resulted in improved agreement. To our knowledge, this is the first time that a particular digital RC has shown to improve agreement among pathologists.
Whole slide digital imaging is studied to see if it can potentially replace traditional microscopy. For example, in a study Ho et al., traditional and whole slide imaging (WSI) methods were found to be comparable when reviewing 24 full genitourinary cases (including 47 surgical parts and 391 slides).  In our case, we determined that the consensus was negatively affected by the WSI. There may be several factors contributing to this result. WSI reading is not commonly done and our pathologists were not used to seeing these. Therefore, human computer interaction and design factors might have played a large role in this. Larger studies with different protocols need to be carried out to further elucidate the reasons.
Improvement in concordance observed for RC3 relative to RC1 can be due to two main factors. First, by enabling pathologists to read exactly the same field, the variability due to the selection of different fields is removed. It is well-known that many tumors contain heterogeneity in cellular distribution and depending on which areas of the slide each pathologist selects, there can be great variation in the average number of CBs noted. Therefore, even if the pathologists are very accurate in their readings, they might be viewing portions of the tissue that reflect different CB counts. The second potential factor is due to the fact that in RC3, errors due to counting are minimized; the CB counting is done on the computer and pathologists have visual cues (i.e., a dot in a marked location) to indicate, which areas of the HPF they have already reviewed and whether a particular cell has already been counted or not. Future studies need to be designed to determine which of these factors play a more important role in improved concordance in pathologists' grading of FL.
The current study suggests a three-phased implementation of a digital reading strategy. In the first phase, well-tested algorithms for the detection of follicles can be used to select 10, random HPFs for the pathologist. By consistently selecting these 10 HPFs, digital reading will improve the concordance of pathologists. In the second phase, these 10 HPFs could be selected by the help of a computer system, which can make sure that the selected fields represent the heterogeneity of the slide. This is expected to reduce the selection bias. In the third phase, detection of CBs in either selected fields or in the whole slide can be carried out with the help of the computer. These detections, can be incorporated in R3 so that pathologists can be presented with cells marked as CB by the computer and/or be given an indication of which grade a particular slide represents according to the computer image analysis. The effect of such systems on the accuracy and concordance need to be determined in future human reader studies.
There were several limitations in our study. First, the number of cases was relatively small. Three different modes of reading were employed, two of which involved digital reading, which is not currently used in clinical practice. In addition none of our readers had prior experience with digital reading. Lack of experience in reviewing digital images combined with the fact that each CB had to be individually marked electronically increased the amount of time each pathologist spent on each case several times more than conventional reading. In our future work, we plan to increase the number of cases and re-assess inter-reader variability among pathologists. Second, all the cases in this study were collected from a single institution with a single method of tissue processing, sectioning and staining. Therefore, these results may or may not be applicable to other cases selected from different institutions. Since the results of our study are comparable to previous studies in conventional reading, we expect this to be a minor limitation. However, future studies will need to include cases from multiple institutions. Third, the selected fields in RC1 could be the same; such an approach would allow us to focus on the digital versus glass comparison. However, for this study's scope such an approach is not practical.
| Acknowledgment|| |
The project described was supported in part by Award Number R01CA134451 from the National Cancer Institute. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Cancer Institute, or the National Institutes of Health.
| References|| |
|1.||Anderson JR, Armitage JO, Weisenburger DD. Epidemiology of the non-Hodgkin's lymphomas: Distributions of the major subtypes differ by geographic locations. Non-Hodgkin's Lymphoma Classification Project. Ann Oncol 1998;9:717-20. |
|2.||López-Guillermo A, Caballero D, Canales M, Provencio M, Rueda A, Salar A, et al. Clinical practice guidelines for first-line/after-relapse treatment of patients with follicular lymphoma. Leuk Lymphoma 2011;52 Suppl 3:1-14. |
|3.||Ardeshna KM, Smith P, Norton A, Hancock BW, Hoskin PJ, MacLennan KA, et al. Long-term effect of a watch and wait policy versus immediate systemic treatment for asymptomatic advanced-stage non-Hodgkin lymphoma: A randomised controlled trial. Lancet 2003;362:516-22. |
|4.||Horning SJ, Rosenberg SA. The natural history of initially untreated low-grade non-Hodgkin's lymphomas. N Engl J Med 1984;311:1471-5. |
|5.||Brice P, Bastion Y, Lepage E, Brousse N, Haïoun C, Moreau P, et al. Comparison in low-tumor-burden follicular lymphomas between an initial no-treatment policy, prednimustine, or interferon alfa: A randomized study from the Groupe d'Etude des Lymphomes Folliculaires. Groupe d'Etude des Lymphomes de l'Adulte. J Clin Oncol 1997;15:1110-7. |
|6.||Colombat P, Salles G, Brousse N, Eftekhari P, Soubeyran P, Delwail V, et al. Rituximab (anti-CD20 monoclonal antibody) as single first-line therapy for patients with follicular lymphoma with a low tumor burden: Clinical and molecular evaluation. Blood 2001;97:101-6. |
|7.||Colombat P, Brousse P, Morschhauser F, Franchi-Rezgui P. Single treatment with rituximab monotherapy for low-tumor burden follicular lymphoma (FL): Survival analyses with extended follow-up of 7 years. Blood 2006;108:147a. |
|8.||National Comprehensive Cancer Network. Clinical Practice Guidelines in Oncology Non-Hodgkin's Lymphoma. Version 2, 2012. |
|9.||Dreyling M, ESMO Guidelines Working Group. Newly diagnosed and relapsed follicular lymphoma: ESMO clinical recommendations for diagnosis, treatment and follow-up. Ann Oncol 2009;20 Suppl 4:119-20. |
|10.||Bierman PJ. Natural history of follicular grade 3 non-Hodgkin's lymphoma. Curr Opin Oncol 2007;19:433. |
|11.||Solal-Céligny P, Roy P, Colombat P, White J, Armitage JO, Arranz-Saez R, et al. Follicular lymphoma international prognostic index. Blood 2004;104:1258-65. |
|12.||Swerdlow SH, Campo E, Harris NL, Jaffe ES, Pileri SA, Stein H, et al. WHO Classification of Tumors of Hematopoietic and Lymphoid Tissue. Geneva: WHO Press; 2008. |
|13.||Mann RB, Berard CW. Criteria for the cytologic subclassification of follicular lymphomas: A proposed alternative method. Hematol Oncol 1983;1:187-92. |
|14.||Metter GE, Nathwani BN, Burke JS, Winberg CD, Mann RB, Barcos M, et al. Morphological subclassification of follicular lymphoma: Variability of diagnoses among hematopathologists, a collaborative study between the Repository Center and Pathology Panel for Lymphoma Clinical Studies. J Clin Oncol 1985;3:25-38. |
|15.||A clinical evaluation of the International Lymphoma Study Group classification of non-Hodgkin's lymphoma. The non-Hodgkin's lymphoma classification project. Blood 1997;89:3909-18. |
|16.||Patil KD. Cochran's Q test: Exact distribution. J Am Stat Assoc 1975;70:186-9. |
|17.||Holm S. A simple sequentially rejective multiple test procedure. Scand J Stat 1979;6:65-70. |
|18.||Landis JR, Koch GG. The measurement of observer agreement for categorical data. Biometrics 1977;33:159-74. |
|19.||Hochberg Y, Tamhane AC. Multiple Comparison Procedures. Vol. 82. New York, NY Wiley Online Library; 1987. |
|20.||Akakin H, Kong H, Elkins C, Hemminger J, Miller B, Ming J, et al. Automated detection of cells from immunohistochemically-stained tissues: Application to Ki-67 nuclei staining. Proceedings of SPIE Medical Imaging Conference, Feb. 4, 2012. Vol. 8315: San Diego, CA; 2012. |
|21.||Belkacem-Boussaid K, Pennell M, Lozanski G, Shana'aah A, Gurcan M. Effect of pathologist agreement on evaluating a computer-assisted system: Recognizing centroblasts in follicular lymphoma cases. Proceedings of IEEE ISBI 2010: Biomedical Imaging from Nano to Macro, p. 1411-4, Rotterdam, The Netherlands, 2010. |
|22.||Belkacem-Boussaid K, Pennell M, Lozanski G, Shana'ah A, Gurcan M. Computer-aided classification of centroblast cells in follicular lymphoma. Anal Quant Cytol Histol 2010;32:254-60. |
|23.||Belkacem-Boussaid K, Prescott J, Lozanski G, Gurcan MN. Segmentation of follicular regions on H and E slides using a matching filter and active contour model. SPIE Medical Imaging 2010: Computer-Aided Diagnosis. Vol. 7624. San Diego CA; 2010. |
|24.||Belkacem-Boussaid K, Samsi S, Lozanski G, Gurcan MN. Automatic detection of follicular regions in H and E images using iterative shape index. Comput Med Imaging Graph 2011;35:592-602. |
|25.||Belkacem-Boussaid K, Sertel O, Lozanski G, Shana'aah A, Gurcan M. Extraction of color features in the spectral domain to recognize centroblasts in histopathology. Conf Proc IEEE Eng Med Biol Soc 2009;2009:3685-8. |
|26.||Cooper L, Sertel O, Kong J, Lozanski G, Huang K, Gurcan M. Feature-based registration of histopathology images with different stains: An application for computerized follicular lymphoma prognosis. Comput Methods Programs Biomed 2009;96:182-92. |
|27.||Samsi S, Lozanski G, Shana'ah A, Krishanmurthy AK, Gurcan MN. Detection of follicles from IHC-stained slides of follicular lymphoma using iterative watershed. IEEE Trans Biomed Eng 2010;57:2609-12. |
|28.||Samsi SS, Krishnamurthy AK, Groseclose M, Caprioli RM, Lozanski G, Gurcan MN. Imaging mass spectrometry analysis for follicular lymphoma grading. Conf Proc IEEE Eng Med Biol Soc 2009;2009:6969-72. |
|29.||Sertel O, Catalyurek UV, Lozanski G, Shanaah A, Gurcan MN. An image analysis approach for detecting malignant cells in digitized H and E-stained histology images of follicular lymphoma. In: 2010 20 th International Conference on Pattern Recognition (ICPR). Istanbul: Turkey; 2010. p. 273-6. |
|30.||Sertel O, Kong J, Catalyurek U, Lozanski G, Saltz J, Gurcan M. Histopathological image analysis using model-based intermediate representations and color texture: Follicular lymphoma grading. J Signal Process Syst 2009;55:169-83. |
|31.||Sertel O, Kong J, Lozanski G, Catalyurek U, Saltz JH, Gurcan MN. Computerized microscopic image analysis of follicular lymphoma. SPIE Medical Imaging 2008: Computer-Aided Diagnosis. Vol. 6915. San Diego: CA; 2008. p. 1-11. |
|32.||Sertel O, Kong J, Lozanski G, Shanaah A, Gewirtz A, Racke F, et al. Computer-assisted grading of follicular lymphoma: High grade differentiation. Mod Pathol 2008;21:371A. |
|33.||Sertel O, Lozanski G, Shana'ah A, Gurcan MN. Computer-aided detection of centroblasts for follicular lymphoma grading using adaptive likelihood-based cell segmentation. IEEE Trans Biomed Eng 2010;57:2613-6. |
|34.||Cambazoglu B, Sertel O, Kong J, Saltz JH, Gurcan MN, Catalyurek UV. Efficient processing of pathological images using the grid: Computer-aided prognosis of neuroblastoma. In: Proceedings of Fifth International Workshop on Challenges of Large Applications in Distributed Environments (CLADE), Monterey Bay, CA. ACM; 2007. p. 35-41. |
|35.||Kumar VS, Kurc T, Kong J, Catalyurek U, Gurcan M, Saltz J. Performance vs. accuracy trade-offs for large-scale image analysis applications. In: 2007 IEEE International Conference on Cluster Computing. Austin: TX; 2007. p. 100-9. |
|36.||Ruiz A, Sertel O, Ujaldon M, Catalyurek U, Saltz J, Gurcan M. Pathological image analysis using the GPU: Stroma classification for neuroblastoma. Proc (IEEE Int Conf Bioinformatics Biomed) 2007;Silicon Valley, CA 78-85. |
|37.||Ruiz A, Kong J, Ujaldon M, Boyer K, Saltz J, Gurcan M. Pathological image segmentation for neuroblastoma using the GPU. In: IEEE ISBI, Paris, France, 2008. p. 296-9. |
|38.||Kumar V, Narayanan S, Kurc T, Kong J, Gurcan M, Saltz J. Analysis and semantic querying in large biomedical image datasets - A set of techniques for analyzing, processing, and querying large biomedical image datasets uses semantic and spatial information. Comput IEEE Comput Mag 2008;41:52-9. |
|39.||Saltz J, Kurc T, Hastings S, Langella S, Oster S, Ervin D, et al. e-Science, caGrid, and Translational Biomedical Research. Computer (Long Beach Calif) 2008;41:58-66. |
|40.||Teodoro G, Sachetto R, Sertel O, Gurcan MN, Meira W, Catalyurek U, et al. Coordinating the use of GPU and CPU for improving performance of compute intensive applications. In: 2009 IEEE International Conference on Cluster Computing and Workshops. New Orleans: LA; 2009. p. 437-46. |
|41.||Patterson ES, Rayo M, Gill C, Gurcan MN. Barriers and facilitators to adoption of soft copy interpretation from the user perspective: Lessons learned from filmless radiology for slideless pathology. J Pathol Inform 2011;2:1. |
|42.||Kong J, Sertel O, Shimada H, Boyer K, Saltz J, Gurcan M. Computer-aided grading of neuroblastic differentiation: Multi-resolution and multi-classifier approach. IEEE Int Conf Image Proc 2007;1-7:2777-80. |
|43.||Kong J, Sertel O, Shimada H, Boyer KL, Saltz JH, Gurcan MN. A multi-resolution image analysis system for computer-assisted grading of neuroblastoma differentiation. SPIE Medical Imaging 2008: Computer-Aided Diagnosis. Vol. 6915. San Diego: CA; 2008. p. 452-60. |
|44.||Kong J, Sertel O, Shimada H, Boyer K, Saltz J, Gurcan M. Computer-aided evaluation of neuroblatoma on whole-slide histology images: Classifying grade of neuroblastic differentiation. Pattern Recognit 2009;42:1080-92. |
|45.||Sertel O, Kong J, Shimada H, Catalyurek UV, Saltz JH, Gurcan MN. Computer-aided prognosis of neuroblastoma on whole-slide images: Classification of stromal development. Pattern Recognit 2009;42:1093-103. |
|46.||Sertel O, Kong J, Shimada H, Catalyurek U, Saltz JH, Gurcan M. Computer-aided prognosis of neuroblastoma: Classification of stromal development on whole-slide images. SPIE Medical Imaging 2008: Computer-Aided Diagnosis. Vol. 6915. San Diego: CA; 2008. p. 44-55. |
|47.||Sertel O, Catalyurek UV, Shimada H, Gurcan MN. Computer-aided prognosis of neuroblastoma: Detection of mitosis and karyorrhexis cells in digitized histological images. Conf Proc IEEE Eng Med Biol Soc 2009;2009:1433-6. |
|48.||Ho J, Parwani AV, Jukic DM, Yagi Y, Anthony L, Gilbertson JR. Use of whole slide imaging in surgical pathology quality assurance: Design and pilot validation studies. Hum Pathol 2006;37:322-31. |
[Figure 1], [Figure 2], [Figure 3], [Figure 4], [Figure 5], [Figure 6]
[Table 1], [Table 2], [Table 3], [Table 4]
|This article has been cited by|
||Informatics Approaches to Address New Challenges in the Classification of Lymphoid Malignancies
| ||Jacob Jordan,Jordan S. Goldstein,David L. Jaye,Metin Gurcan,Christopher R. Flowers,Lee A.D. Cooper |
| ||JCO Clinical Cancer Informatics. 2018; (2): 1 |
|[Pubmed] | [DOI]|
||The Pathologist 2.0: An Update on Digital Pathology in Veterinary Medicine
| ||Christof A. Bertram,Robert Klopfleisch |
| ||Veterinary Pathology. 2017; 54(5): 756 |
|[Pubmed] | [DOI]|
||Clinical, pathological and genetic features of follicular lymphoma grade 3A: a joint analysis of the German low-grade and high-grade lymphoma study groups GLSG and DSHNHL
| ||K. Koch,E. Hoster,M. Ziepert,M. Unterhalt,G. Ott,A. Rosenwald,M. L. Hansmann,W. Bernd,H. Stein,V. Pöschel,M. Dreyling,L. Trümper,M. Löffler,N. Schmitz,W. Hiddemann,M. Pfreundschuh,W. Klapper |
| ||Annals of Oncology. 2016; : mdw185 |
|[Pubmed] | [DOI]|
||Assessment of pathological response to therapy using lipid mass spectrometry imaging
| ||Nathan Heath Patterson,Balqis Alabdulkarim,Anthoula Lazaris,Aurélien Thomas,Mieczyslaw M. Marcinkiewicz,Zu-hua Gao,Peter B. Vermeulen,Pierre Chaurand,Peter Metrakos |
| ||Scientific Reports. 2016; 6: 36814 |
|[Pubmed] | [DOI]|