14

Improving Prostate Cancer Care through Quality Assurance Programs

Hans Veerman1,2,3* Marinus J. Hagens1,2,3* André N. Vis2,3 R. Jeroen A. van Moorselaar2,3 Pim J. van Leeuwen1,3 Michel W.J.M. Wouters4,5,6 Henk G. van der Poel1,2,3

1Department of Urology, Netherlands Cancer Institute – Antoni van Leeuwenhoek Hospital, Amsterdam, The Netherlands; 2Department of Urology, Amsterdam University Medical Centers location Boelelaan, Amsterdam, The Netherlands; 3Prostate Cancer Network the Netherlands, Amsterdam, The Netherlands; 4Scientific Bureau, Dutch Institute for Clinical Auditing, Leiden, The Netherlands; 5Department of Biomedical Data Sciences, Leiden University Medical Center, Leiden, The Netherlands; 6Department of Medical Oncology, Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital, Amsterdam, The Netherlands

Abstract: Continuous quality assurance assessment and control in healthcare is essential to provide patients with the best possible care. Quality assurance programs have been developed to improve future healthcare by thoroughly studying patient outcomes on a physician- or institutional-level. Through the continuous and cyclical process of data registration, evaluation and adaptation, opportunities are sought to improve (individual) patient outcomes. Over the past decade, quality assurance programs have been initiated within urological clinical practice, mainly focusing on the diagnosis and surgical treatment of prostate cancer. While they all share the same philosophy to improve healthcare, existing quality assurance programs differ greatly. To date, little is known about their effects on the outcomes of prostate cancer care. In this chapter, we summarize the current knowledge regarding quality assurance program within prostate cancer care. We provide insights into how quality Assurance programs can improve and assure future diagnosis and treatment of prostate cancer.

Keywords: cyclical quality assurance for prostate cancer; improving prostate cancer care; quality assurance for prostate cancer; requirements for quality assurance programs; statistical quality assurance for prostate cancer

Author for correspondence: Hans Veerman, Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital (NCI-AVL), Plesmanlaan 121, 1066 CX Amsterdam, The Netherlands. Email: h.veerman@nki.nl

Cite this chapter as: Veerman H, Hagens MJ, Vis AN, van Moorselaar RJA, van Leeuwen PJ, Wouters MWJM, van der Poel HG. Improving Prostate Cancer Care through Quality Assurance Programs. In: Barber N and Ali A, editors. Urologic Cancers. Brisbane (AU): Exon Publications. ISBN: 978-0-6453320-5-6. Online first 26 May 2022.

Doi: https://doi.org/10.36255/exon-publications-urologic-cancers-prostate-cancer-care

In: Barber N, Ali A (Editors). Urologic Cancers. Exon Publications, Brisbane, Australia. ISBN: 978-0-6453320-5-6. Doi: https://doi.org/10.36255/exon-publications-urologic-cancers

Copyright: The Authors.

License: This open access article is licenced under Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) https://creativecommons.org/licenses/by-nc/4.0/

*These authors contributed equally to the creation of this manuscript

INTRODUCTION

Quality assurance and control has rapidly become part of the everyday vocabulary in healthcare. In the early 1900s, with his landmark publication, Ernest Amory Codman was pioneering in this field (1, 2). With the End Results System, Codman advocated tracking patient outcomes in order to improve the quality of healthcare. It was his belief that high-quality care did not derive from fancy equipment, but rather from self-assessment by healthcare professionals. Although being ostracized by his colleagues for his idea, it forms the basis of many contemporary initiatives to improve the quality of healthcare worldwide.

Following in Codman’s footsteps, physicians have sought a more scholarly approach to quality assurance in healthcare by acquiring knowledge and expertise from the industrial sector (3, 4). Prospective registries, the contemporary, more advanced equivalent to Codman’s End Results Cards, have been implemented and play an important role in present-day quality assurance programs (QAPs). QAPs are structured programs in which healthcare employees critically review the outcomes of their patients and continuously analyze and discuss these results in order to improve the outcomes.

Healthcare QAPs originated in general surgery, but over the past decade they have also been initiated in the urological practice with a particular focus on the diagnosis and surgical treatment of prostate cancer (PCa) (5). The formation and structure of these QAPs have previously been described, but little is known about the effects of these QAPs on the outcomes of PCa care (6). Therefore, this chapter reviews the available literature on QAPs in PCa care, answering the following questions: (ii) what is the theory behind QAPs; (ii) which organizational requirements are necessary; and (iii) what is the available evidence on the effect of QAPs on PCa care?

CYCLICAL QUALITY ASSURANCE

QAPs use continuous and short-cycled processes of data registration, evaluation, and adaptation to improve outcomes. This ideology did not originate from healthcare, but from the production industry. In the 1950s, Dr. William Edwards Deming developed a cyclical technique to address and solve problems in production lines and thereby improve the quality of industrial/organizational processes continuously, herewith building on the work of Dr Walter Andrew Shewhart (the Plan, Do, Check, Act (PDCA)-cycle; Figure 1) (7). Deming’s philosophy revolutionized the industrial output in post-war Japan, where this philosophy is known as “kaizen”—the continuous search for opportunities for all processes to get better (3). Although Deming’s PDCA-cycle was intended for the industrial sector, it could also be applied in healthcare systems (8). The PDCA-cycle is comprised of four steps: (Plan) identifying clinical steps that require improvement; (Do) implementing interventions; (Check) evaluating clinical outcomes after these interventions; and (Act) implementing these interventions (if outcomes are favorable) in clinical practice.

Fig 1

Figure 1. The Plan, Do, Check, Act (PDCA)-cycle (or Deming’s cycle) – a continuous and cyclical technique to improve outcomes. Figure from: https://www.praxisframework.org/en/library/shewhart-cycle

After the completion of one full cycle, a new period of data collection, data analysis and evaluation ensue. The length of each cycle depends on what is being investigated at the time; a sufficient number of events must occur to detect a change in outcome. Therefore, cycle lengths can range from three cycles in 1 day to one cycle in 16 months (9). Depending on the objective to be achieved, the duration of a cycle chain (first to last cycle of one chain) may also vary enormously (1 day to 4 years). To manage the duration and analyze the quality, it is imperative to have a predetermined end date of the cycle. Therefore, before starting the PDCA-cycle, a statistical well-designed power calculation is essential.

STATISTICAL QUALITY ASSURANCE

It is essential to analyze outcomes in a correct manner to carefully target improvement efforts and assess the success of implemented pathways, protocols, and improvement plans (10). The use of statistical process control (SPC), developed by the aforementioned Dr. Shewhart, aids in testing the effectiveness of an intervention. SPC has found its way into healthcare systems over the last two decades (11).

In every (production) process, two types of variation can be distinguished: common cause variation and special cause variation, both affecting (product) quality (10, 12). Common cause variation is defined as variance inherent to the process itself; similar to many population characteristics that follow a Gaussian distribution with approximately 5% of measurements that fall outside of the 2 standard deviation limits. Special cause variation, on the other hand, is defined as variance that can be attributed to a specific cause (i.e., an intervention). The presence of special cause variation is a signal that the process has changed (either for better or for worse). Shewhart developed run and control charts to distinguish these types of variation within a production line process. For illustrative purposes, a control chart was created indicating the number of prostate biopsies for a hypothetical cohort of men suspected of having prostate cancer (Figure 2).

Fig 2

Figure 2. Control chart representing the average number of transperineal prostate biopsies per patient over time in a teaching hospital. Normal cause variation is present due to differences in physicians and baseline characteristics of patients; however, special cause variation was observed between July and October 2021. Special cause variation was defined as any outcome above or below 3 standard deviations (SDs) and 4 out of last 5 outcomes above or below 1 SD. This substantial rule violation was accompanied by the introduction of a new physician; a physician-in-training took significantly more prostate biopsies per patient. Through performance feedback and discussion, the number of biopsies normalized again as of November 2021.

Although effective in detecting large shifts in a production line, Shewhart control charts are unable to find moderate or small shifts. This reduced sensitivity can be compensated for by augmenting Shewhart control charts with cumulative sum (CUSUM) control charts (13, 14). Unlike Shewhart control charts, CUSUM control charts represent information of current and previous samples at each point. Plotting the cumulative sums of deviations from the target value of current and previous samples results in greater sensitivity for detecting shifts or trends over the traditional Shewhart control charts. By way of example, a CUSUM chart was created indicating the number of positive surgical margins of one surgeon’s consecutively treated patients (Figure 3).

Fig 3

Figure 3. Risk-adjusted cumulative sum (CUSUM) plot of hypothetical data. The plot represents the surgical margin status of one surgeon’s consecutively treated patients. First, a prediction model is created using logistic regression with clinical variables as input and positive surgical margins (PSM) as output. This model predicts the probability of PSM for each individual patient. The probability ranges from 0 to 1. If a patient had a surgical margin (unwanted outcome), a score of 1 minus the predicted probability is added to the cumulative sum (line goes up). If a patient had a negative surgical margin (desired outcome) the predicted probability is subtracted from the cumulative sum (line goes down). The control limit is calculated using the standard deviation of the mean proportion of PSM and the group size and a weighted parameter (usually 4). If a small difference should be detected or if there are little data points, the weighted parameter can be decreased. For indicative purposes, the control limit is set at 3. The control limit is reached at the 94th patient, which indicates that the surgeon had more PSM than predicted based on the patients’ clinical variables. This could be a reason to evaluate the surgeon’s technique to improve the PSM rate.

In addition to assessing the success of implemented interventions, comparing outcomes of physicians/hospitals is an important feature of QAPs; not to stimulate competition, but to identify variation in care processes that may be associated with outcomes (outcomes research). However, given that patient populations may differ between physicians and hospitals, assessment of and adjustment for case-mix variation is warranted. Methods have been proposed to analyze whether physicians/hospitals differ in outcome when compared to the mean risk of the case-mix subgroup (15, 16). By using multivariable regression models, the observed/expected ratio (O/E ratio) of each physician/hospital and outcome can be calculated. This is the for case-mix adjusted ratio indicating the quality of a physician/hospital. As a visual aid, case-mix adjusted funnel plots can be constructed; any physician/hospital that falls outside of the 95% confidence intervals has outcomes that significantly deviate from the average of the group. Figure 4 shows a funnel plot of hypothetical data illustrating the O/E ratios of positive surgical margins of 9 different surgeons.

Fig 4

Figure 4. A funnel plot was constructed of a hypothetical cohort of prostate cancer patients who underwent robot-assisted radical prostatectomy. The funnel plot displays the observed/expected (O/E)-ratio of positive surgical margins (PSM) of 9 different surgeons. One surgeon, highlighted as 1, is depicted above the 95% confidence interval upper control limit, which indicates that they make more PSM than expected based on the patients’ clinical characteristics when compared with the other surgeons. This could be a reason to evaluate the technique of surgeon 1 in order to improve the PSM rate.

REQUIREMENTS FOR QAPS

Properly constructing a QAP is essential for its success. Given the continuous nature of QAPs, the first and foremost requirement is motivated physicians that form the steering committee of the program. Their indispensable input identifies clinical steps or outcomes that require improvement. Subsequently, these outcomes must be properly registered in prospective (institutional) databases. Quality assurance can be performed both on a physician or institutional level. Multicenter collaborations and/or hospital networks facilitate an inter-institutional comparison. However, the accompanying data transfers between hospitals can be problematic because of technical or (patient)privacy issues. These problems may be solved by using secured internet-based, multicentric electronic data capture (EDC) systems managed by an independent data processor, who takes care of pseudonomization of patient level data before analyses are performed (17).

Before starting data collection, consensus must be reached on relevant outcomes indicative of quality of care. The International Consortium for Health Outcomes Measurement (ICHOM) has developed specific sets of relevant patient outcome measures for both localized and advanced PCa, that can be registered in a standardized way (18, 19). Confidence in data accuracy and completeness is fundamental for quality assurance and for the provision of sufficiently robust evidence on which to base changes in practice recommendations. Physician-reported data is reliable to benchmark outcomes (20). However, the data collection process should be well described and monitored in order to provide accurate and complete data that can be used for multiple purposes (i.e., research or hospital management). To prevent bias, both analyses and interpretations should be performed by independent parties. Additionally, participants put themselves in a vulnerable position in which they receive feedback (criticism) on their professional functioning, therefore data should be handled confidentially. Presenting the data in a safe environment in which the participants can freely discuss without repercussions is a prerequisite. This can be achieved by anonymizing data or limiting data access to participants of the QAPs only (21). Participating physicians/hospitals are expected to trust the data and the data collection process; if they perceive the feedback as non-credible, they may not be motivated to change their practice.

The last requirement is back-up by hospital managers. Drafting, implementing, and maintaining a QAP requires monetary investment. Therefore, the functioning of the QAP must also be evaluated: is it cost-effective? Depending on the subject, calculating quality-adjusted life years (QALYs) or the incremental cost-effectiveness ratio (ICER), defined by the difference in cost between two interventions divided by the difference in their effect, can provide insight into its cost-effectiveness. Improvement in the quality of care is associated with less comorbidity and less frequent follow-up treatment. Therefore, QAPs can lead to long-term cost reduction (2224).

EFFECTS OF QAPS IN PCA CARE

We reviewed the literature on studies assessing the effect if QAPs in PCa care. In this, we specifically searched for studies that mention a QAP or improvement cycle according to the definitions of section 2.

The first attempts to develop QAPs for PCa care were made in Sweden. The merger of several regional databases created the National Prostate Cancer Registry (NPCR) (25). Using this database, surgeons were able to compare outcomes of their hospital to historical data of other hospitals and to the national average. In 2017, the NPCR opted for full transparency. All outcomes were made publicly available through an online dashboard in order to stimulate national quality control. With the help of this online dashboard, physicians can compare hospital-specific outcomes between Swedish hospitals. To preserve the privacy of physicians, individual surgeon-specific outcomes are only accessible to colleagues within their own department. The NPCR has already proven its worth. In 2014, an increased rate of readmissions after prostatectomy was observed, mainly due to anastomotic leaks. Videos of these patients were reviewed, and the literature was searched to identify surgical steps during the apical dissection and suturing the anastomosis. As a result, the surgical technique was changed, which led to a decrease in the readmission rate (from 10.6% to 5%) (25).

The Michigan Urology Statewide Improvement Collaboration (MUSIC) is a group of 46 urology practices and over 250 participating urologists (26). The QAP of MUSIC aims to improve the quality and cost-efficiency of PCa care, by reducing variance in practice. Within MUSIC, participating urologists submit data to a web-based clinical registry and, subsequently, receive quarterly reports in which their performance is compared to the statewide average and to other physicians. To date, they have published several papers on quality assurance in both PCa diagnosis and treatment. MUSIC underscores the positive changes that can be achieved with the collaborative QAP on PCa diagnostics. In an effort to improve data completeness, MUSIC has shown that QAPs are able to improve documentation of key variables, such as the clinical TNM-classification. By educating a dedicated urologist in each participating center on the importance of clinical TNM-classification for clinical decision-making and having them share this and their performance data with other members of their practice, documentation ultimately improved (27). Through performance feedback and education interventions, imaging appropriateness has been improved and biopsy-related complications have been reduced (2729). Additionally, MUSIC has focused on the variation in surgeon-specific outcomes, such as erectile dysfunction, urinary incontinence, and complicated postoperative recovery. In accordance with the NPCR, MUSIC argues that objective identification of surgeons who achieve better outcomes will provide insight into specific techniques associated with those better outcomes (3032). It has been suggested that peer reviewing of surgical videos and coaching may improve surgical skills and, hopefully, patient outcomes (33, 34).

Participating in a nationwide QAP is mandatory in Germany. In 2008, the German Cancer Society (Deutschen Krebsgesellschaft (DKG)) initiated a certification program to increase the quality of PCa care in Germany (35). To qualify for certification, centers must have established a quality management system and meet quality indicators yearly. Fifteen quality indicators (both treatment and process related) were established based on expert opinion and clinical guidelines. Despite the efforts made, improvements in functional and oncological outcomes could not be demonstrated (36). In the meantime, the German Martini Clinic implemented a physician-initiated QAP on its own initiative. They realized that their institutional prospective database, initially started for scientific purposes, could also be utilized to aid in a QAP. This data collection contributed significantly to constant quality improvements over the years (37). For example, anesthetic regimens were adapted, which led to a decrease in intraoperative blood loss. In addition, nomograms were implemented and the NeuroSAFE technique was developed, which increased nerve-sparing procedures while keeping biochemical recurrence rates steady (38). They also noticed that one of their urologists had improved urinary continence outcomes compared to the others. After watching surgical videos, they found that the surgeon used a specific technique when dissecting the prostatic apex and urethra. Implementation of this technique by all other surgeons improved the urinary continence rate of all surgeons (39).

The London Cancer Network noticed poorer results compared to international colleagues, which motivated them to initiate a QAP. Through image-based surgical planning and monthly peer reviewing of individual surgeons’ outcomes, a high quality of care for patients undergoing radical prostatectomy was pursued (40). The implementation of such a QAP substantially improved quality of care, in terms of both oncological and functional outcomes; nerve-sparing surgery increased significantly while margin status remained static, and postoperative urinary continence and erectile function improved.

Similar to the MUSIC approach, Veerman et al. aimed to reduce catheter-related bladder discomfort (CRBD) after robotic-assisted radical prostatectomy by applying a QAP to the intra-operative anesthesia regime (41). After 8 cycles of different treatments and adapting the treatment protocol, the optimal treatment regime was identified. This regime reduced the incidence of CRBD from 70% to 36%, a relative reduction of 49%. Matulewicz et al. sought to determine the efficacy of comparative quality performance review to improve a surgeon-level measure of surgical oncologic quality. Participating surgeons were provided with confidential report cards detailing information about their patients’ clinical characteristics and positive surgical margin rates (42). These report card also contained information on their historical data, the institutional average, and the blinded results of peers. Before implementation of report cards, the positive surgical margin rate was 10.6%, while during and after the implementation, the positive surgical margin rate dropped significantly to 7.4%.

DISCUSSION

QAPs are increasingly implemented to achieve and maintain high quality PCa care. The currently existing QAPs differ in focus, execution, motivation, and subjects; however, they all share the same philosophy: to improve future PCa care by thoroughly studying their own retrospective data and identifying outliers. The existing literature has already described interesting results of QAPs on PCa care, such as improvement of functional outcomes, improvement of oncological outcomes and reduction of variability between physicians/hospitals.

QAPs are no stand-alone “research” projects. They are continuous cycles, incorporated into daily practice, striving for the best possible care. While the quality cycles warrant the quality of care, participating physicians should warrant the quality of the cycle. Quality assurance is achieved though the collection and analysis of reliable data and the willingness of physicians to act on these findings. Maintaining prospective registry databases alone is insufficient. Comparing one’s results with peers or the hospital average gives a good indication of performance, but to improve outcomes, discussion between peers and identifying improvement steps is essential. The willingness to improve must come from the physicians themselves; physician-initiated programs have been shown to improve both functional and oncological outcomes, whereas programs in which physicians were enforced to participate did not improve outcomes (35).

No consensus exists on the level of data transparency; opinions differ on the accessibility of quality assurance data. The Swedes have opted for full transparency and have made their data publicly available through an interactive online dashboard. Some, however, believe that patients may misinterpret the outcomes by a lack of context and lack of medical or epidemiological knowledge. Consequently, patients may avoid physicians and/or hospitals with a ‘worse’ performance. Besides, full data transparency may evoke risk-aversive behavior in physicians by not treating high-risk patients, induce registration bias, and limit physicians’ motivation. These actions are counterproductive for the progressive nature of the QAP (43, 44). On the other hand, full or partial access to quality assurance data for professionals is more accepted and even beneficial. Shared insight into the data improves physicians’ confidence in the data accuracy. Moreover, physicians can benchmark their results to the average and to their peers. Consequently, participants can identify points of improvement and they find solutions through a joint approach. In this way, the participants can learn from each other. Transparency of data within selected groups is therefore recommended to maximize the positive effects of QAP.

Criticism on the current literature on QAPs is that its focus is on improving one outcome at a time. Associations with multiple outcomes are not always taken into account. PCa care is dynamic and important outcomes (urinary incontinence, erectile functioning, positive surgical margin, etc.) are related to each other. An improvement of a specific outcome does not necessarily mean an improvement of the whole; other outcomes may be adversely affected by the intervention. The London Cancer Network and the German Martini Clinic should be commended in this respect as they reported all relevant outcomes instead of focusing on only one single quality indicator.

In order to obtain reliable results, a high volume of treated cases is desirable. High-volume centres generally have better outcomes than low-volume centres (45, 46). Additionally, improving patient outcomes through short cycles of quality improvement is easier with a higher volume of treated patients. After all, when a low volume of patients is treated, it may take years to measure a difference in outcomes after a change of practice. Collaborations between hospitals, such as networks and forming a hub-and-spoke model, can effectuate a high patient volume. The formation of hospital networks offers several other advantages. For example, centralisation is associated with increased compliance to guidelines and reduced costs. Centralisation also offers novel surgeons the opportunity to learn from expert surgeons, which may increase the quality of care in the entire network (47, 48).

Randomised controlled trials (RCTs) are the purest way to demonstrate a causal relationship between intervention and effect. However, there is a major drawback to this research method. Due to the highly selected populations used in RCTs, outcomes in daily practice may differ from RCT results. ‘Real world” insights, gained through QAPs, can therefore be of great value in complementing evidence from RCTs (49). In the case of QAPs an unselected population is used, so outcomes are more based on daily practice and reliable for physicians. In addition, high costs are involved in carrying out RCTs, whereas QAPs are associated with minimal costs (50).

Many papers that publish on quality improvement initiatives in PCa care have positive outcomes. This may indicate a publication bias. Centres with less appealing results may be afraid to publish or struggle to find journals that accept their research. Consequently, it is harder to attribute the trend in improved results to the QAP alone. Furthermore, all centres that published on the effects of QAPs on PCa care are high-volume centres that are actively involved in the scientific community. It is possible that the effect is caused by applying the latest scientific insights, rather than learning from the best surgeon in the group. However, the results of the Dutch Institute of Clinical Auditing (DICA) counter this argument. DICA performs quality cycles regarding the treatment of several (non-urological) oncological and non-oncological diseases. All hospitals that treat a specific condition participate in the corresponding quality cycle, making it a nation-wide, population-based QAP. Results of the DICA QAPs have shown several improvements in quality of care, i.e., improved reporting, decreased between-hospital variations, decreased complication rates and improved mortality rates (51).

CONCLUSION

In conclusion, despite differences in organizational characteristics, the available literature shows positive effect of QAP (providing that motivated participants are involved). The use of QAP should therefore be recommended in urological practices. The key for success is a group of motivated physicians who lead the QAP.

Conflict of Interest: The authors declare no potential conflicts of interest with respect to research, authorship and/or publication of this chapter.

Copyright and Permission Statement: The authors confirm that the materials included in this chapter do not violate copyright laws. Where relevant, appropriate permissions have been obtained from the original copyright holder(s), and all original sources have been appropriately acknowledged or referenced.

REFERENCES

  1. Codman EA. A Study in Hospital Efficiency: As Demonstrated by the Case Report of the First Five Years of a Private Hospital. Boston: Thomas Todd Co.; 1918
  2. Rodkey G, Itani K. Evaluation of healthcare quality: a tale of three giants. Am J Surg. 2009;198(5):S3–S8. https://doi.org/10.1016/j.amjsurg.2009.08.004
  3. Berwick D. Continuous Improvement as an Ideal in Health Care. N Engl J Med. 1989;320(1):53–56. https://doi.org/10.1056/NEJM198901053200110
  4. Donabedian A. The quality of care. How can it be assessed?. JAMA. 1988;260(12):1743–1748. https://doi.org/10.1001/jama.1988.03410120089033
  5. Gaylis F, Nasseri R, Salmasi A, Anderson C, Mohedin S, Prime R et al. Implementing Continuous Quality Improvement in an Integrated Community Urology Practice: Lessons Learned. Urology. 2021;153:139–146. https://doi.org/10.1016/j.urology.2020.11.068
  6. Reese A, Ginzburg S. The past, present, and future of urological quality improvement collaboratives. Transl Androl Urol. 2021;10(5):2280–2288. https://doi.org/10.21037/tau.2019.10.18
  7. Aguayo R. Dr. Deming: The American Who Taught the Japanese About Quality. New York: Fireside Simon and Schuster; 1991.
  8. Schneider PD. FOCUS-PDCA ensures continuous quality improvement in the outpatient setting. Oncol Nurs Forum. 1997;24:966.
  9. Taylor M, McNicholas C, Nicolay C, Darzi A, Bell D, Reed J. Systematic review of the application of the plan-do-study-act method to improve quality in healthcare. BMJ Qual Saf. 2013;23(4):290–298. https://doi.org/10.1136/bmjqs-2013-001862
  10. Carey R. How Do You Know That Your Care Is Improving? Part I. J Ambul Care Manage. 2002;25(1):80–87. https://doi.org/10.1097/00004479-200201000-00011
  11. Shewhart WA. Economic control of quality of manufactured product. London: Macmillan And Co Ltd: 1931.
  12. Carey R. How Do You Know That Your Care Is Improving? Part II. J Ambul Care Manage. 2002;25(2):78–88. https://doi.org/10.1097/00004479-200204000-00010
  13. Noyez L. Shewhart control charts and the cumulative sum test - the use of clinical data in quality control. Ned Tijdschr Geneesk. 2010;154:A1245.
  14. Grigg O, Farewell V, Spiegelhalter D. Use of risk-adjusted CUSUM and RSPRT charts for monitoring in medical contexts. Stat Methods Med Res. 2003;12(2):147–170. https://doi.org/10.1177/096228020301200205
  15. Elfrink A, van Zwet E, Swijnenburg R, den Dulk M, van den Boezem P, Mieog J et al. Case-mix adjustment to compare nationwide hospital performances after resection of colorectal liver metastases. Eur J Surg Oncol. 2021;47(3):649–659. https://doi.org/10.1016/j.ejso.2020.10.016
  16. Lawson K, Daignault K, Abouassaly R, Khanna A, Martin L, Goldenberg M et al. Hospital-level Effects Contribute to Variations in Prostate Cancer Quality of Care. Eur Urol Oncol. 2021;4(3):494–497. https://doi.org/10.1016/j.euo.2020.08.010
  17. Hernandez-Boussard T, Blayney D, Brooks J. Leveraging Digital Data to Inform and Improve Quality Cancer Care. Cancer Epidemiol Biomarkers Prev. 2020;29(4):816–822. https://doi.org/10.1158/1055-9965.EPI-19-0873
  18. Martin N, Massey L, Stowell C, Bangma C, Briganti A, Bill-Axelson A et al. Defining a Standard Set of Patient-centered Outcomes for Men with Localized Prostate Cancer. Eur Urol. 2015;67(3):460–467. https://doi.org/10.1016/j.eururo.2014.08.075
  19. Morgans A, van Bommel A, Stowell C, Abrahm J, Basch E, Bekelman J et al. Development of a Standardized Set of Patient-centered Outcomes for Advanced Prostate Cancer: An International Effort for a Unified Approach. Eur Urol. 2015;68(5):891–898. https://doi.org/10.1016/j.eururo.2015.06.007
  20. Aning J, Parry M, Meulen J, Fowler S, Payne H, McGrath J et al. How reliable are surgeon-reported data? A comparison of the British Association of Urological Surgeons radical prostatectomy audit with the National Prostate Cancer Audit Hospital Episode Statistics-linked database. BJU Int. 2021;128(4):482–489. https://doi.org/10.1111/bju.15399
  21. Vickers A, Sjoberg D, Basch E, Sculli F, Shouery M, Laudone V et al. How Do You Know If You Are Any Good? A Surgeon Performance Feedback System for the Outcomes of Radical Prostatectomy. Eur Urol. 2012;61(2):284–289. https://doi.org/10.1016/j.eururo.2011.10.039
  22. Aggarwal A, Aeran H, Rathee M. Quality management in healthcare: The pivotal desideratum. J Oral Biol Craniofac Res. 2019;9(2):180–182. https://doi.org/10.1016/j.jobcr.2018.06.006
  23. Cheung K, Moiduddin A, Chin M, Drum M, Brown S, Graber J et al. The Perceived Financial Impact of Quality Improvement Efforts in Community Health Centers. J Ambul Care Manage. 2008;31(2):111–119. https://doi.org/10.1097/01.JAC.0000314701.50042.0b
  24. Govaert J, van Bommel A, van Dijk W, van Leersum N, Tollenaar R, Wouters M. Reducing Healthcare Costs Facilitated by Surgical Auditing: A Systematic Review. World J Surg. 2015;39(7):1672–1680. https://doi.org/10.1007/s00268-015-3005-9
  25. Stranne J, Axen E, Franck-Lissbrant I, Fransson P, Frånlund M, Hugosson J et al. Single institution followed by national implementation of systematic surgical quality control and feedback for radical prostatectomy: a 20-year journey. World J Urol. 2019;38(6):1397–1411. https://doi.org/10.1007/s00345-019-02887-4
  26. Montie J, Linsell S, Miller D. Quality of Care in Urology and the Michigan Urological Surgery Improvement Collaborative. Urol Pract. 2014;1(2):74–78. https://doi.org/10.1016/j.urpr.2014.04.003
  27. Filson C, Boer B, Curry J, Linsell S, Ye Z, Montie J et al. Improvement in Clinical TNM Staging Documentation Within a Prostate Cancer Quality Improvement Collaborative. Urology. 2014;83(4):781–787. https://doi.org/10.1016/j.urology.2013.11.040
  28. Auffenberg G, Qi J, Gao Y, Miller D, Ye Z, Brachulis A et al. Evaluation of a needle disinfectant technique to reduce infection-related hospitalisation after transrectal prostate biopsy. BJU Int. 2017;121(2):232–238. https://doi.org/10.1111/bju.13982
  29. Hurley P, Dhir A, Gao Y, Drabik B, Lim K, Curry J et al. A Statewide Intervention Improves Appropriate Imaging in Localized Prostate Cancer. J Urol. 2017;197(5):1222–1228. https://doi.org/10.1016/j.juro.2016.11.098
  30. Myers S, Ghani K, Dunn R, Lane B, Schervish E, Gao Y et al. Notable Outcomes and Trackable Events after Surgery: Evaluating an Uncomplicated Recovery after Radical Prostatectomy. J Urol. 2016;196(2):399–404. https://doi.org/10.1016/j.juro.2016.02.083
  31. Auffenberg G, Qi J, Dunn R, Linsell S, Kim T, Miller D et al. Evaluation of Patient- and Surgeon-Specific Variations in Patient-Reported Urinary Outcomes 3 Months After Radical Prostatectomy From a Statewide Improvement Collaborative. JAMA Surg. 2021;156(3):e206359. https://doi.org/10.1001/jamasurg.2020.6359
  32. Agochukwu-Mmonu N, Qi J, Dunn R, Montie J, Wittmann D, Miller D et al. Patient- and Surgeon-Level Variation in Patient-Reported Sexual Function Outcomes Following Radical Prostatectomy Over 2 Years. JAMA Surg. 2022;157(2):136. https://doi.org/10.1001/jamasurg.2021.6215
  33. Prebay Z, Peabody J, Miller D, Ghani K. Video review for measuring and improving skill in urological surgery. Nat Rev Urol. 2019;16(4):261–267. https://doi.org/10.1038/s41585-018-0138-2
  34. Ghani K, Miller D, Linsell S, Brachulis A, Lane B, Sarle R et al. Measuring to Improve: Peer and Crowd-sourced Assessments of Technical Skill with Robot-assisted Radical Prostatectomy. Eur Urol. 2016;69(4):547–550. https://doi.org/10.1016/j.eururo.2015.11.028
  35. Kowalski C, Ferencz J, Albers P, Fichtner J, Wiegel T, Feick G et al. Quality assessment in prostate cancer centers certified by the German Cancer Society. World J Urol. 2015;34(5):665–672. https://doi.org/10.1007/s00345-015-1688-z
  36. Schlomm T, Huland H, Graefen M. Improving Outcome of Surgical Procedures Is Not Possible Without Adequate Quality Measurement. Eur Urol. 2014;65(6):1017–1019. https://doi.org/10.1016/j.eururo.2013.11.042
  37. Steuber T, Graefen M, Chun K, Haese A, Schlomm T, Erbersdobler A et al. 1575: Validation of a Nomogram for Prediction of Side-Specific Extracapsular Extension at Radical Prostatectomy. J Urol. 2006;175(4S):507–508. https://doi.org/10.1016/S0022-5347(18)33767-4
  38. Schlomm T, Tennstedt P, Huxhold C, Steuber T, Salomon G, Michl U et al. Neurovascular Structure-adjacent Frozen-section Examination (NeuroSAFE) Increases Nerve-sparing Frequency and Reduces Positive Surgical Margins in Open and Robot-assisted Laparoscopic Radical Prostatectomy: Experience After 11 069 Consecutive Patients. Eur Urol. 2012;62(2):333–340. https://doi.org/10.1016/j.eururo.2012.04.057
  39. Schlomm T, Heinzer H, Steuber T, Salomon G, Engel O, Michl U et al. Full Functional-Length Urethral Sphincter Preservation During Radical Prostatectomy. Eur Urol. 2011;60(2):320–329. https://doi.org/10.1016/j.eururo.2011.02.040
  40. Cathcart P, Sridhara A, Ramachandran N, Briggs T, Nathan S, Kelly J. Achieving Quality Assurance of Prostate Cancer Surgery During Reorganisation of Cancer Services. Eur Urol. 2015;68(1):22–29. https://doi.org/10.1016/j.eururo.2015.02.028
  41. Veerman H, Houwink A, Schutte P, Nieuwenhuijzen J, Roeleveld T, Wit E et al. Intraoperative Strategies to Reduce Catheter-Related Bladder Discomfort in the Early Postoperative Period after Robot-Assisted Radical Prostatectomy. J Urol. 2021;205(6):1671–1680. https://doi.org/10.1097/JU.0000000000001645
  42. Matulewicz R, Tosoian J, Stimson C, Ross A, Chappidi M, Lotan T et al. Implementation of a Surgeon-Level Comparative Quality Performance Review to Improve Positive Surgical Margin Rates during Radical Prostatectomy. J Urol. 2017;197(5):1245–1250. https://doi.org/10.1016/j.juro.2016.11.102
  43. Moscucci M, Eagle K, Share D, DeFranco A, Vakili B, Clark V et al. Public reporting and case selection for percutaneous coronary interventions: an analysis from two large multicenter PCI databases. J Am Coll Cardiol. 2002;39:436. https://doi.org/10.1016/S0735-1097(02)81957-X
  44. Narins C, Dozier A, Ling F, Zareba W. The Influence of Public Reporting of Outcome Data on Medical Decision Making by Physicians. Arch Intern Med. 2005;165(1):83. https://doi.org/10.1001/archinte.165.1.83
  45. Nathan A, Gershman B, Van der Poel H, Sooriakumaran P. Centralisation of Care for Prevalent Urological Malignancies: The Case for Prostate Cancer. Eur Urol Focus. 2021;7(5):920–923. https://doi.org/10.1016/j.euf.2021.08.012
  46. Tilki D, Huland H, Graefen M. Centralization and Quality Control of Elective Surgery Improve Outcome: Aren’t We Ethically Obliged to Force the Pace of Creating High-volume Centers?. Eur Urol. 2015;68(1):30–31. https://doi.org/10.1016/j.eururo.2015.03.032
  47. Reitblat C, Bain P, Porter M, Bernstein D, Feeley T, Graefen M et al. Value-Based Healthcare in Urology: A Collaborative Review. Eur Urol. 2021;79(5):571–585. https://doi.org/10.1016/j.eururo.2020.12.008
  48. Gershman B, Meier S, Jeffery M, Moreira D, Tollefson M, Kim S et al. Redefining and Contextualizing the Hospital Volume-Outcome Relationship for Robot-Assisted Radical Prostatectomy: Implications for Centralization of Care. J Urol. 2017;198(1):92–99. https://doi.org/10.1016/j.juro.2017.01.067
  49. Hoeijmakers F, Beck N, Wouters M, Prins H, Steup W. National quality registries: how to improve the quality of data?. J Thorac Dis. 2018;10(S29):S3490-S3499. https://doi.org/10.21037/jtd.2018.04.146
  50. Speich B, von Niederhäusern B, Schur N, Hemkens L, Fürst T, Bhatnagar N et al. Systematic review on costs and resource use of randomized clinical trials shows a lack of transparent and comprehensive data. J Clin Epidemiol. 2018;96:1–11. https://doi.org/10.1016/j.jclinepi.2017.12.018
  51. Beck N, van Bommel A, Eddes E, van Leersum N, Tollenaar R, Wouters M. The Dutch Institute for Clinical Auditing: Achieving Codman’s Dream on a Nationwide Basis. Ann Surg. 2020;271(4):627–631. https://doi.org/10.1097/SLA.0000000000003665