We use cookies to help you navigate efficiently and perform certain functions. You will find detailed information about all cookies under each consent category below.
The cookies that are categorized as "Necessary" are stored on your browser as they are essential for enabling the basic functionalities of the site....
Necessary cookies are required to enable the basic features of this site, such as providing secure log-in or adjusting your consent preferences. These cookies do not store any personally identifiable data.
Functional cookies help perform certain functionalities like sharing the content of the website on social media platforms, collecting feedback, and other third-party features.
Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics such as the number of visitors, bounce rate, traffic source, etc.
Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.
Advertisement cookies are used to provide visitors with customized advertisements based on the pages you visited previously and to analyze the effectiveness of the ad campaigns.
From finding better ways to develop exam content to discovering innovative methods for scoring competency-based assessments, NBME is exploring how technology can be used to solve the most pressing challenges facing the medical education community.
Creating a scalable approach to generating high-quality exam content
As the health care landscape continues to evolve and more emphasis is placed on the formative assessment of physician competencies, the need for up-to-date and relevant exam content is more important than ever. Traditionally, the process of developing high-quality and reliable questions has not only required significant time and resources but is also not easily scalable.
This created a unique challenge for medical educators and one that NBME sought to address by exploring the application of automated item generation (AIG) in medical education assessment. AIG involves the use of cognitive modeling to automatically generate a high volume of multiple-choice questions, enabling educators to access a wider selection of content to measure student knowledge.
After running an initial proof-of-concept pilot in early 2020, NBME trained 38 subject matter experts from our test development committees on creating AIG models, which led to the generation of questions that were pretested on four NBME® Clinical Science Subject Exams and the Health & Wellness Coach Certifying Exam. Item analyses indicate the AIG questions are performing at least as well as traditionally written items.
The success of this pilot inspired the creation of a new AIG tool specially tailored to NBME test items called IMAGE (Item Modeling and Automated Generation Engine), which is currently being used to generate questions for additional NBME exams.
Developing a more efficient and transparent process for scoring clinical text
From speech recognition software to translation apps that foster better communication, Natural Language Processing (NLP) has made remarkable advances over the past decade. NBME is exploring new ways that this technology can be applied to medical education assessment.
One such application involves using NLP to automate the scoring of clinical text, such as patient notes or short-answer questions, with the goal of making the process more transparent and efficient. As outlined by Victoria Yaneva, Senior Data Scientist at NBME, this poses a unique research challenge:
“Having physicians score clinical text requires significant time, along with human and financial resources. NLP can aid this process by automatically mapping concepts from a scoring rubric to concepts found in the clinical text written by an examinee. However, this mapping is not always straightforward.
Some concepts such as ‘loss of interest in activities’ can be expressed as ‘no longer plays tennis,’ while other concepts may require combining multiple text segments or dealing with ambiguous linguistic constructions as in ‘no cold intolerance, hair loss, palpitations, or tremor’ corresponding to ‘lack of other thyroid symptoms.’ These cases require creative computational solutions.”
To help advance the area of clinical text scoring, NBME sponsored a Kaggle competition on this topic. Data scientists from all over the world enrolled to develop the best NLP scoring approach using the same dataset of clinical text. The May 2022 results will remain open source, so others can benefit and continue developing shared value for the community.