Radiologia Brasileira - Publicação Científica Oficial do Colégio Brasileiro de Radiologia

AMB - Associação Médica Brasileira CNA - Comissão Nacional de Acreditação
Idioma/Language: Português Inglês

Vol. 51 nº 3 - May / June  of 2018

ORIGINAL ARTICLE
Print 

Page(s) 172 to 177

Reproducibility of a quantitative system for assessing the quality of diagnostic ultrasound

Autho(rs): Wagner Iared1; Andrea Puchnick2; Eduardo Bancovsky3; Paulo Roberto Bettini3; Leonardo Modesti Vedolin4; Maria Cristina Chammas5

PDF English      PDF Português

Texto em Português English Text

Keywords: Ultrasonography; Quality control; Quality improvement; Certification; Accreditation.

Descritores: Ultrassonografia; Controle de qualidade; Melhoria de qualidade; Certificação; Acreditação.

Abstract:
OBJECTIVE: To present a quantitative system for assessing the quality of ultrasound examinations—SQUALUS—and to determine its reproducibility, taking into consideration the images on file, as well as the consistency between the images obtained and the final report.
MATERIALS AND METHODS: The system includes questions related to the number of images; the appropriateness of images in relation to the protocol established; focus adjustment; depth; gain; and appropriateness of the measurements for B-mode examinations. For Doppler examinations, the system includes questions related to the appropriateness of color images, the spectral analysis, and correction of the insonation angle. To assess the quality of the report, the system includes questions related to the consistency between the images obtained and the contents of the report. An overall numerical score was assigned by averaging the scores for image quality and for the contents of the report. Two independent examiners, each blinded to the evaluation of the other, assessed 30 different types of ultrasound examinations.
RESULTS: There was statistically significant agreement between the two examiners for 8 of the 10 questions related to image quality. For the questions related to the quality of the reports, the interexaminer agreement was almost perfect.
CONCLUSION: The proposed quantitative system for assessing the quality of ultrasound examinations is a reproducible tool that can be used in audits and accreditation programs.

Resumo:
OBJETIVO: Apresentar e verificar a reprodutibilidade de um sistema quantitativo de avaliação da qualidade de exames de ultrassonografia (SQUALUS), levando em consideração as imagens documentadas e a coerência entre as imagens obtidas e o laudo final.
MATERIAIS E MÉTODOS: Foram elaborados quesitos considerando o número de imagens, a adequação das imagens ao protocolo estabelecido, o ajuste do foco, a profundidade e ganho e a adequação das medidas. Para exames com Doppler também foram avaliadas a adequação das fotos coloridas, a análise espectral e a correção do ângulo. Para a qualidade do laudo foi considerada a coerência com as imagens documentadas e seu conteúdo. Um sistema numérico foi atribuído conferindo uma nota final à qualidade das imagens, ao conteúdo do laudo e à média das duas avaliações. Trinta exames de ultrassonografia de diferentes tipos foram avaliados por dois examinadores independentes, cegos à avaliação um do outro.
MATERIAIS E MÉTODOS: Os avaliadores apresentaram concordância estatisticamente significante em 8 de 10 quesitos para avaliação da qualidade da imagem. Na avaliação dos laudos, a concordância entre os avaliadores foi quase perfeita.
CONCLUSÃO: O sistema quantitativo de avaliação da qualidade de exames ultrassonográficos proposta é uma ferramenta reprodutível que pode ser utilizada em auditorias e em programas de acreditação.

INTRODUCTION

There is real concern about the quality of medical services in Brazil. The National Health Insurance Agency of the Brazilian National Ministry of Health has developed norms and standards, such as the Program for the Dissemination of Health Insurance Provider Qualifications, established by Normative Resolution no. 267 of August 24, 2011, and the Program for Monitoring the Quality of Health Insurance Providers, established by Normative Resolution no. 275 of November 1, 2011(1). Those programs promote the quality of providers focused on beneficiaries. The programs are also in line with the efforts of the Colégio Brasileiro de Radiologia e Diagnóstico por Imagem (CBR, Brazilian College of Radiology and Diagnostic Imaging) in its Programa de Acreditação em Diagnósticos por Imagem (PADI, Program for the Accreditation of Diagnostic Imaging Clinics), as well as in its quality certification programs(2,3).

Quality assessments are performed differently for ultrasound examinations than for other diagnostic imaging examinations. In modalities such as mammography, radiography, computed tomography, and magnetic resonance imaging, non-medical professionals (technicians, technologists, and biomedical professionals) acquire and register images according to certain protocols, which can be altered on a case-by-case basis by radiologists. In ultrasound, however, it is the primary care physician who acquires the images and prepares the final report. Although the CBR, through the Comissão Nacional de Ultrassonografia (CNUS, National Ultrasound Commission) and the PADI(2,3), together with other institutions, has standardized protocols for minimum documentation of the images obtained in each type of ultrasound examination, the results vary markedly among different physicians and facilities.

With respect to the ultrasound report, unlike other imaging modalities, in which the same images can be interpreted by more than one radiologist, ultrasound is a dynamic method in which the scanning of the various organs and structures must be performed in several planes, although only a limited number of images are recorded in the documentation. It is presumed that the images recorded are the most representative of the normal and pathological findings, according to what is specified in the minimum documentation protocols. This makes it difficult to evaluate the technical quality of ultrasound examinations.

To date, there have been no studies suggesting any method of evaluating the quality of ultrasound examinations similar to the one we are proposing. Although there have been studies focusing on the technical aspects of the equipment(4,5), the technical parameters related to training methods(6), and the influence of image quality on specific diagnoses(7–9), there have been none focusing on the quality of the examination performed and reported by medical professionals.

The objective of this work is to present and determine the reproducibility of a quantitative system of quality evaluation of ultrasound examinations—the Sistema Quantitativo de Avaliação da Qualidade de Exames de Ultrassonografia (SQUALUS, Quantitative System for Assessing the Quality of Ultrasound Examinations)—taking into account the images on file and the concordance between the images obtained and the final report.


MATERIALS AND METHODS

Development of criteria for evaluation of image quality


To evaluate the examinations, we used a checklist based on the PADI and CNUS/CBR regulations(2,3), taking into consideration the technical parameters proposed by the American College of Radiology and the American Institute of Ultrasound in Medicine(10).

For routine examinations, the following parameters were used:

• Minimum recommended number of images.
• Structures documented in accordance with the recommendations.
• Appropriate depth.
• Focus adjustment.
• Adjustment of transducer gain/frequency/gray-scale map
• Check of the appropriateness of the pertinent measures.

For Doppler examinations, the following were also evaluated:

• Appropriateness of color Doppler images.
• Appropriate spectral analysis.
• Angle correction for the analysis of flow velocities.

For each item, the evaluator should check and answer yes, no, or not applicable. For different types of examination, different weights were checked for each item. Table 1 shows the checklist used for all types of examinations, and Table 2 shows the weights of each question for the different types of examination. If the answer was yes for all questions, the total score would be 10. If an examination was totally non-standard—that is, the answer was no for all questions—the total score would be 0.







Development of criteria for evaluating the appropriateness of the report and its concordance with the images

Because imaging is operator-dependent, the final report of the ultrasound examination was evaluated according to its concordance with the images obtained. The degree of agreement between the report and the images was classified into four different categories:

• The report is appropriate and consistent with the images.
• There is disagreement, but the disagreement is of little relevance.
• There is disagreement of moderate relevance.
• There is highly relevant disagreement.

Minor typographical errors and spelling mistakes that did not compromise the understanding of the report and the final diagnosis were classified as inconsistencies of little relevance. Gross errors in the text, inappropriate descriptions, and dubious diagnostic impressions were classified as moderately relevant disagreements. Descriptions and diagnostic impressions inconsistent with the images on file were classified as highly relevant disagreements. Appropriate, consistent reports were given a score of 10; reports containing inconsistencies of little relevance were given a score of 7; reports with moderately relevant disagreements were given a score of 3; and reports with highly relevant disagreements were given a score of 0.

The report and the images are inseparable, so the arithmetic mean between the quality score assigned to the images and the score for the appropriateness/consistency of the report was used as the final parameter to judge the quality of each examination evaluated. If an examination presented excellent image quality and the report had highly relevant disagreements, the final combined score was 5. If the images were excellent and the report had moderately relevant disagreements, the final combined score was 6.5. If the report was appropriate and consistent but the images did not meet any of the recommended criteria, the final combined score was 5.

Concordance test

Prior to performing the examinations, the medical team was informed of the documentation protocol to be followed and was made aware of the fact that the examinations would be randomly audited. The identity of the physicians who performed the examinations involved in this study, as well as the identity of the patients, was kept confidential.

Two independent evaluators, both CBR-certified radiologists, with 23 and 16 years of experience in ultrasound, respectively, evaluated the same 5 examinations of each type, for a total of 30 evaluations per evaluator. The examinations evaluated were randomly extracted from those performed in the laboratories of a private company in São Paulo, between August and September of 2016. Each evaluator applied the checklist and was blinded to the evaluation of the other. The types of examinations evaluated were as follows: total abdominal ultrasound; transvaginal pelvic ultrasound; ultrasound of the breasts; Doppler ultrasound of thyroid nodules; Doppler ultrasound of the thyroid for evaluation of diffuse disease; Doppler ultrasound of the carotid artery; and Doppler ultrasound of the veins of the lower limbs.

Statistical analysis

The data collected in this study were initially submitted to descriptive statistical analysis. For the quantitative (continuous) variables "image score", "report score", and "mean final examination score", we calculated summary measures, such as mean and standard deviation. Qualitative (categorical) variables were analyzed by calculating absolute and relative frequencies.

For all of the parameters evaluated, the agreement between the two evaluators was quantified by calculating the kappa statistic (k). To quantify the agreement between the scores and means obtained by the two evaluators, a Bland–Altman plot was constructed, after which the intraclass correlation coefficient (kicc) was estimated. For all measurements, 95% confidence intervals were calculated. The strength of the agreement, based on the kappa coefficients (k;kicc), was interpreted as shown in Table 3.




Statistical evaluation was performed with the Statistical Package for the Social Sciences, version 16.0 for Windows (SPSS Inc., Chicago, IL, USA). For all conclusions reached by inferential analysis, the level of significance was set at 5%.


RESULTS

The evaluators presented statistically significant agreement for 8 of the 10 questions (Table 4). The strength of agreement ranged from perfect to poor.




In relation to the reproducibility of the image scores, report scores, and final means, the differences between the analyses of evaluator 1 and those of evaluator 2 are shown in the Bland–Altman plot (Figure 1). Overall, in the three parameters analyzed, the evaluators presented very similar values (mean differences varying between −0.1 ± 0.9 and −0.1 ± 1.2), indicating almost perfect agreement (Table 5).


Figure 1. Distribution of the scores for the images and the reports, together with the final mean scores, given for the examinations by evaluators 1 and 2. The mean of the difference between the two evaluators was –0.1 ± 1.2 in relation to the image scores (A), –0.1 ± 1.1 in relation to the report scores (B) and –0.1 ± 0.9 compared with the final mean examination score (C). Black line: mean of the differences. Gray lines: 95% confidence intervals.




DISCUSSION

The almost perfect agreement between the evaluators for the mean image quality scores, the report scores, and the final mean scores of the examinations demonstrates that the method is reproducible. For the checklist criteria with the greatest relevance for the final mean score, the agreement was classified as significant and perfect. Although depth, focus on the center of the structure, and gain were less concordant, these data did not significantly affect the final mean score, because each of those criteria had a weight of only one point in the mean for all types of examinations.

The ideal form of evaluation of the quality of ultrasound examinations would be through a retrospective analysis, correlating their results with the clinical follow-up and eventual findings on other imaging tests, as well as with surgical and pathological findings. Such an evaluation method may be applicable in the hospital environment, where integration between radiologists and teams of other specialties facilitates the monitoring and favorable evolution of cases. The examinations performed on an outpatient basis do not have the benefit of that type of confirmation, being limited to the evaluation of the person who performs them.

The quality of the photographic documentation of ultrasound examinations, following established documentation protocols, should not be considered mere whim, knowing that in some cases excellent physicians perform accurate diagnoses with the method, although without recording the established patterns. The observation of such protocols, in addition to serving as an eventual legal support, proves that the examination was carried out, exhausting, from a technical point of view, the method in question. Examinations with poorly documented images are classified as poor quality examinations.

Adequate photographic documentation demonstrates that a refined technique was used, which confers significantly greater sensitivity on the method, proving that scans were performed in different planes of the studied organs. The adjustment of depth and focus optimizes the detection of lesions and the detailing of their characteristics.

The evaluated parameters cover the number of images, the appropriateness of the images to the established protocol, the care in the adjustment of the focus, depth and gain, and adequacy of the measurements, for the examinations in mode B. For Doppler examinations, we evaluated the appropriateness of the color photos, spectral analysis, and angle correction. In the SQUALUS method, the first two parameters evaluated are precisely to determine whether the examination followed the minimum documentation guidelines. The weight for each criterion was assigned in a consensus meeting of the authors, aiming at valuing the essential aspects of each type of examination. That was the most time-consuming phase of the study.

The final quantitative result allows the accrediting entity to define a cut-off point from which the examination is considered acceptable. Our suggestion is that examinations with a final mean score of 7 or higher be considered qualified, because the reports are consistent or contain only inconsistencies of little relevance, which do not imply disagreement with the final diagnosis. With scores 0 and 3 for disagreements of high or moderate relevance, respectively, these cases would never reach a mean score of 7, even if the image quality score was 10. Likewise, appropriate reports with an image quality score lower than 4 would not reach the mean score of 7. In our experience, in poorly documented examinations, the evaluators rarely find good agreement between the images and the report. However, we must emphasize, that this study is limited by the small number of examinations of each type, which made it impossible to assess agreement for each of them.

For examinations performed in urgent and emergency situations, limited documentation or even a report without documented images is acceptable. However, without proper recording of images, it is impossible for the quality of the final result to be validated through audits conducted by the facility itself or by accreditation programs. Therefore, when submitting examinations to accreditation programs, care should be taken to follow the minimum protocols of photographic documentation required.


CONCLUSION

The SQUALUS is a reproducible tool that can be used in audits and accreditation programs.


REFERENCES

1. Agência Nacional de Saúde Suplementar. Ministério da Saúde. Resolução Normativa - RN nº 275, de 1º de novembro de 2011. Brasília: ANS/MS; 2011. [cited 2017 Jan 8]. Available from: http://bvsms.saude.gov.br/bvs/saudelegis/ans/2011/res0275_01_11_2011.html.

2. Colégio Brasileiro de Radiologia e Diagnóstico por Imagem. Selos CBR. São Paulo: CBR; 2015. [cited 2017 Jan 8]. Available from: https://cbr.org.br/programas-de-qualidade/selos-cbr/.

3. Colégio Brasileiro de Radiologia e Diagnóstico por Imagem. Programa de Acreditação em Diagnóstico por Imagem. São Paulo: CBR; 2017. [cited 2017 Jan 8]. Available from: http://padi.org.br/apresentacao/.

4. Welsh D, Inglis S, Pye SD. Detecting failed elements on phased array ultrasound transducers using the Edinburgh Pipe Phantom. Ultrasound. 2016;24:68–93.

5. Trier HG. Quality assurance in ultrasound diagnostic equipment in West Germany. Experiences with current regulations. Ultraschall Med. 1985;6:255–64.

6. Schäfer RD. Quality assurance of Doppler sonographic procedures—educational concepts of medical programs. Ultraschall Med. 1985;6:279–81.

7. Dhombres F, Roux N, Friszer S, et al. Relation between the quality of the ultrasound image acquisition and the precision of the measurement of the crown-rump length in the late first trimester: what are the consequences? Eur J Obstet Gynecol Reprod Biol. 2016;207:37–44.

8. Choi JI, Jung SE, Jeong WK, et al. Effectiveness of on-site education for quality assurance of screening ultrasonography for hepatocellular carcinoma. Med Ultrason. 2016;18:275–80.

9. Silverman NH, Moon-Grady AJ. Discordant fetal and post-natal diagnosis: can we do better? J Am Coll Cardiol. 2016;68:931–3.

10. American College of Radiology. Practice parameter for performing and interpreting diagnostic ultrasound examinations. ACR–SPR–SRU Res. 7 – 2011, Amended 2014 (Res. 39). [cited 2017 Jan 8]. Available from: https://www.acr.org/Quality-Safety/Standards-Guidelines/Practice-Guidelines-by-Modality/Ultrasound.










1. MD, PhD, Radiologist, Advisor for the Graduate Program in Evidence-Based Medicine of the Escola Paulista de Medicina da Universidade Federal de São Paulo (EPM-Unifesp), São Paulo, SP, Brazil
2. Professor, Coordinator of Instruction and Research in the Department of Diagnostic Imaging of the Escola Paulista de Medicina da Universidade Federal de São Paulo (EPM-Unifesp), São Paulo, SP, Brazil
3. MD, Radiologist, Coordinator of the Ultrasound Division of Diagnósticos da América S/A, Barueri, SP, Brazil
4. MD, PhD, Neuroradiologist, Director of Radiology and Diagnostic Imaging at Diagnósticos da América S/A, Barueri, SP, Brazil
5. MD, PhD, Radiologist, Director of the Ultrasound Department of the Instituto de Radiologia do Hospital das Clínicas da Faculdade de Medicina da Universidade de São Paulo (InRad/HC-FMUSP), São Paulo, SP, Brazil

Study conducted in the Department of Diagnostic Imaging of the Escola Paulista de Medicina da Universidade Federal de São Paulo (EPM-Unifesp), São Paulo, SP, Brazil.

Mailing address:
Dr. Wagner Iared
Departamento de Diagnóstico por Imagem – EPM-Unifesp
Rua Napoleão de Barros, 800, Vila Clementino
São Paulo, SP, Brazil, 04024-002
E-mail: wagneriared@gmail.com

Received February 11, 2017.
Accepted after revision May 8, 2017.
 
RB RB RB
GN1© Copyright 2024 - All rights reserved to Colégio Brasileiro de Radiologia e Diagnóstico por Imagem
Av. Paulista, 37 - 7° andar - Conj. 71 - CEP 01311-902 - São Paulo - SP - Brazil - Phone: (11) 3372-4544 - Fax: (11) 3372-4554