Revolutionizing orthodontic treatment planning with AI and relational reasoning
Millions of people around the world undergo some type of orthodontic treatment each year due largely to developmental deformities in the jaw, skull, or face. Computed tomography (CT) imaging is the go-to technique for surgeons when planning such treatments, especially surgeries. This is because CT provides 3D images of the bones and teeth, which helps the surgeon analyze complex cases in detail and determine the best treatment procedure based on that.
During the CT scan, surgeons typically try to pinpoint specific anatomical landmarks in the images. These are distinct points in the human body that can be used as a reference to make measurements and assess a condition or deformity. However, finding these landmarks can be time-consuming and requires considerable skill. Many researchers have therefore attempted to automate this process with artificial intelligence (AI), achieving varying levels of success.
A common problem with existing AI approaches is that they rely on a process known as “segmentation.” In medical image analysis, segmentation implies separating an image into different relevant regions, such as individual bones or specific tissue groups. While this approach works well enough for most people, they tend to fail for patients with implants or deformities, including missing or broken bones. But what if we could find a way to perform anatomical landmarking without needing to segment the image first?
This was the goal of a study conducted by a research team from Northwestern University. The researchers hypothesized that a deep learning AI model should be able to learn the spatial relationships among the anatomical landmarks of the craniomaxillofacial (CMF) bones (bones of the skull, face, and jaw) without requiring an explicit image segmentation. “This approach where an AI model can automatically learn the relationships between anatomies and their underlying reasons is known as “relational reasoning”. While well-known in robotics, relational reasoning has hardly been considered in medical imaging,” explains corresponding author Ulas Bagci, Associate Professor at Northwestern University’s Radiology and Biomedical Engineering Department.
When designing the model, the researchers sought to answer the following questions: (a) Is it possible to identify all anatomical landmarks based on learning only a few of them? (b) Which landmarks are the most informative for the model? An important aspect of their strategy was to implement a model architecture that could learn both “local” and “global” relations. A local relation refers to the relative position between a pair of landmarks, whereas a global relation refers to the position of a landmark in relation to those of all other landmarks.
Accordingly, the architecture of the proposed AI model, which they called a “relational reasoning network” (RRN), has two stages incorporating core blocks known as “relational units.” In the first stage, the model learns about local relations between the members of a given set of landmarks. In the second stage, the model learns about global relations between each landmark and the rest. The team trained the model with a large dataset of landmarks derived from an artificially augmented 250-image dataset of CT scans. A good portion of the patients included in the dataset presented with birth defects, developmental deformities, missing bones or teeth, and previous surgical interventions.
The researchers tested several combinations of landmarks to determine the combination with best performance. They also compared the model’s performance to conventional AI-based landmarking methods. Overall, the accuracy of the RRN was remarkable, in line with or better than that of previously reported techniques. Moreover, the model showed good generalizability, meaning that it performed well when tested with previously unseen data gathered in different conditions. The researchers suggest that this was because the RRN framework could learn the functional relationships between CMF landmarks that are still present to some degree in cases of large deformities.
“With an error less than 2 mm per anatomical landmark in the most difficult cases, the method developed in our study could not only help surgeons save time but also avoid incorrect landmarking that might arise from segmentation failures. Moreover, it sets a precedent for future AI models aimed at learning relationships between anatomical landmarks in other parts of the body,” concludes Bagci.