Submitted Successfully!
Thank you for your contribution! You can also upload a video entry or images related to this topic.
Check Note
2000/2000
Ver. Summary Created by Modification Content Size Created at Operation
1 -- 3516 2022-05-20 10:23:57 |
2 format change -2 word(s) 3514 2022-05-20 12:27:31 |
Knee Injury Detection Using Deep Learning
Edit
Upload a video

Knee injuries account for the largest percentage of sport-related, severe injuries (i.e., injuries that cause more than 21 days of missed sport participation). The improved treatment of knee injuries critically relies on having an accurate and cost-effective detection. Deep-learning-based approaches have monopolized knee injury detection in MRI studies.

ACL deep learning knee injury machine learning meniscus
Information
Contributors MDPI registered users' name will be linked to their SciProfiles pages. To register with us, please refer to https://encyclopedia.pub/register : , , , , ,
View Times: 358
Revisions: 2 times (View History)
Update Date: 20 May 2022
Table of Contents

    1. Introduction

    1.1. Backdrop

    Knee injuries account for the largest percentage of sport-related, severe injuries (i.e., injuries that cause more than 21 days of missed sport participation) [1][2][3][4]. Anterior cruciate ligament (ACL) ruptures represent more than 50% of the cases, affecting 200,000 individuals in the United States each year [1][5][6][7]. Knee cartilage lesions affect around 900,000 individuals in the United States every year, resulting in over 200,000 surgical procedures [5][6][7][8]. Menisci injuries are the second most common knee impairment, with an incidence of 12–14% [9] and a prevalence of 60–70 cases per 100,000 in the United Kingdom [2]. ACL injuries alone account for an expenditure of more than $7 billion in the United States [10]. Both short- and long-term pain, disability, and negatively affected, health-related quality of life have all been strongly associated with knee injuries [11][12][13]. In regard to young and athletic individuals, the more time they spend engaging in occupational and/or recreational activities, the higher predisposition to knee injuries they have, which, in turn, contributes to a higher likelihood of developing osteoarthritis (OA) [14]. On average, half of the individuals, that have an injury that involved ACL and/or meniscal tear develop radiographically confirmed knee OA ten to 20 years post-injury [15][16]. Another two possible consequences of knee injuries are: (i) structural muscle injuries of the lower limb [17]; and (ii) tendinopathies [18]. All the above reflect the direct and indirect (lost wages, productivity, and disability) socio-economic burden conferred on the society by knee injuries. The high prevalence of knee injuries in the general population, and the resulting socio-economic impact, have created a necessity for developing accurate and cost-effective procedures that can detect and quantify the severity of knee injuries. Early diagnosis and, consequently, treatment of ligament rupture, menisci tear, and/or cartilage lesion can prevent early onset of knee OA [1].
    Arthroscopy is considered the “gold-standard” for the diagnosis of intra-articular knee pathologies, but is limited by potential complications and its invasive nature [19]. Therefore, magnetic resonance imaging (MRI) is the most widely used, non-invasive imaging technique for diagnosing knee injuries [20][21]. However, the MRI-based diagnosis of knee injuries can be a very challenging procedure, with the experience of clinicians playing a critical role in image interpretation. Human-based image interpretation pitfalls, such as subjectivity, distraction, and fatigue, as well as diagnostic uncertainties, often lead to erratic diagnoses, hindering the optimal management of knee injuries [22][23]. Moreover, clinical-diagnostic discrepancies among non-musculoskeletal radiologists and orthopedic surgeons are commonly encountered in everyday clinical practice [11].
    Due to the above-listed factors, as well as the exponentially increasing number of clinical examinations, the idea of using computers for improving the challenging task of image interpretation of medical examinations has been recently adopted by the scientific community [24]. Imaging data proliferation, algorithmic advances, and recent technological advances in fast computing have already resulted in a strong push towards the utilization of artificial intelligence (AI) algorithms in medical image analysis. The term AI broadly refers to any method that enables computers to mimic human intelligence [25]. Deep learning (DL) in particular is a class of machine-learning (ML) algorithms that is currently driving the AI boom [26]. Numerous applications of DL in medical image analysis have been reported, including skin cancer classification, diabetic retinopathy detection, lung nodule detection, and mammography cancer detection, among others [27]. The aforementioned AI-empowered solutions are expected to revolutionize medical sectors by improving the accuracy and productivity of different diagnostic and therapeutic measures in clinical practice [20].
    Drawing attention to the diagnosis of knee injuries, several early DL studies have exhibited better performance than traditional ML techniques, while in some cases they have proved to be even superior to radiologists [26]. However, the previously published review studies in the MRI field were either focused on other application domains (e.g., fracture detection [28]) or limited to the performance of the proposed networks without paying attention to their specifics (learning methodology, processing stages, technical limitations etc.) [29].

    1.2. Machine Learning in a Nutshell: Definitions and Terminology

    To enhance the understanding of the readers and for the sake of completeness, this section quickly presents the relevant terminology and definitions with respect to ML and DL algorithms used in the studies involved here. ML is a branch of AI that focuses on the development of algorithms that automatically learn to make accurate predictions by relying on experience (data) rather than on hard-coded instructions.
    Supervised ML systems (Figure 1) operate in two phases: the learning phase (training) and the testing one. In a traditional ML pipeline, a feature extraction/selection stage (also referred to as feature engineering) is first implemented to extract or identify the most informative features [16]. These features can be extracted from the input images, employing various algorithms including grey-level co-occurrence matrix (GLCM), first- and second-order statistics, and shape/edge features, among others [30]. Next, a ML model is fit to the extracted features and the optimal model parameters are obtained. During the testing phase, the trained model is shown previously unseen samples (represented as images or features extracted from images), which are then classified. As opposed to traditional programming, where the rules are manually crafted by a programmer, a supervised ML algorithm automatically formulates rules from the data.
    Figure 1. Examples of typical machine-learning and deep-learning pipelines.
    DL [31] is a subfield of ML that sets an alternative architectural paradigm by shifting the process of extracting features from images to the underlying learning mechanism. The most informative features for the task at hand are extracted by the algorithm itself. The mainstream DL architecture for computer vision applications is the convolutional neural network (CNN). A CNN typically consists of multiple building blocks (layers such as convolutional, pooling, and fully connected) that automatically extract increasingly abstract spatial hierarchies of features. The CNN training is carried out via a backpropagation algorithm. The huge popularity of CNNs is attributed to certain characteristics they possess, such as weight sharing and spatial invariance.
    Transfer learning is a common strategy where a network, that was pre-trained on a big dataset, is partly re-used to provide decisions on a problem with a different dataset. The main idea behind transfer learning is that generic features learned on a large dataset could be useful and applicable to other domain tasks with a potentially limited amount of accessible data. Numerous pre-trained networks are currently available, such as DenseNet [32], AlexNet [33], and VGG [34]. When employing DL with transfer learning for feature extraction, the pre-trained network is treated as an arbitrary feature extractor: the input image propagates through multiple layers until it reaches a pre-specified layer, the outputs of which are considered as the finally extracted features (Table 1).
    Table 1. Brief presentation of the feature extraction techniques, as well as the ML and DL models, and the main procedures.

    Category

    Models

    Description

    Feature extraction

    Histogram of oriented gradient (HOG) [35]

    This is a feature descriptor used in computer vision and image processing for the purpose of object detection. The technique counts occurrences of gradient orientation in localized portions of an image.

    Generalized search tree (GIST) [30]

    GIST descriptor represents holistic spatial scene properties (spatial envelope) of an image. It summarizes gradient information on different spatial scales and orientations by splitting the image into a grid of cells on several scales and convolving each cell using a Gabor filter bank from different perspectives.

    Gray-level co-occurrence matrix (GLCM) [36]

    GLCM is a way of extracting second-order statistical texture features. In particular, the texture of an image is estimated by calculating how often pairs of pixels with specific values and a certain spatial relationship occur.

    Traditional Machine Learning

    k-nearest neighbor (K-NN) [37]

    KNN algorithm is a simple, easy-to-implement supervised ML algorithm that can be used to solve both classification and regression problems. It works by (i) finding the distances between a query and all the examples in the data, (ii) selecting the K nearest neighbors of the query, and (iii) voting for the most frequent label (in the case of classification) or averaging the labels (in the case of regression).

    Support vector machines (SVMs) [38]

    SVMs is a supervised method that identifies a hyperplane that best divides the data into two classes. To separate the two clouds of data points, there are many possible hyperplanes that could be chosen. The objective of the SVM algorithm is to find a slab that has the maximum thickness, i.e., the maximum distance between data points of the different classes.

    Shallow artificial neural networks (ANNs) [39]

    The ANN vaguely simulates the way the human brain analyzes and processes information. They consist of sequential layers: input, hidden and output layers. The hidden layer processes and transmits the input information to the output layer.

    Deep Learning

    Convolutional neural networks (CNNs) [40]

    This is a class of DL algorithms commonly used in computer vision and pattern recognition. CNNs are a specific type of neural networks that are generally composed of the following layers: (i) input layer, (ii) convolution layers, (iii) pooling layers and (iv) fully connected layers. The convolution layers use filters that perform convolution operations as they are scanning the input with respect to its dimensions. Pooling is a down-sampling operation, which is typically applied after a convolution layer. The fully connected layers operate on a flattened input where each input is connected to all neurons in the next layer and are usually found towards the end of CNN architectures to optimize objectives such as class scores.

    Region based convolutional neural networks (R-CNNs) [41]

    The method of detecting and classifying objects in an image is known as object detection. R-CNN (regions with convolutional neural networks) is a deep learning technique that blends rectangular area proposals with convolutional neural network functionality. The R-CNN algorithm is a two-stage detection method.

    Deep residual networks [42]

    A residual neural network (ResNet) is an ANN variant that uses residual mapping and shortcut connections to tackle the problem of vanishing and exploding gradients that is characteristic of deep CNNs. As a consequence of this, deep residual networks achieve better performance when compared to plain very deep networks, whereas their training is easier as well. Typical ResNet models are implemented with double- or triple-layer skips that contain nonlinearities such as rectified linear unit (ReLUs) and batch normalization in between.

    3D-CNNs [43]

    A 3D CNN is simply the 3D generalization of 2D CNNs. It takes as input a 3D volume or a sequence of 2D frames (e.g., slices in an MRI scan). Then kernels move through 3 dimensions of data producing 3D activation maps. Overall, they learn powerful representations of volumetric data.

     

    Computer Vision Transformers [44]

    When data is modelized as a sequence of embeddings, the Transformer model is a basic yet scalable technique that can be used for any type of data. Even without typical convolutional pipelines, transformers can be utilized to provide SOTA results in Computer Vision. It is a DL network that extracts inherent properties of the interest domain via the self-attention technique.

    Procedure

    Training

    The standard procedure involves a dataset of paired images and labels (x, y) for training and testing, an optimizer (e.g., stochastic gradient descent, Adam [45]), and a loss function to update the model parameters. The aim of the training is to find the optimal values for the network parameters so that the loss function is minimized.

    Data augmentation

    Data augmentation is a strategy that artificially generates more training samples to increase the diversity of the training data. This can be done via applying affine transformations (e.g., rotation, scaling), flipping or cropping to original labeled samples.

    Dropout

    Dropout is a regularization method that randomly drops some units from the neural network during training, encouraging the network to learn a sparse representation. It is used to reduce overfitting.

    Loss function

    The metric to assess the discrepancy between model predictions and labels is called loss function. The gradients of the loss function are used to update the weights of the neural networks.

    Transfer learning

    This aims to transfer knowledge from one task to another different but related target task. This is often achieved by reusing the weights of a pre-trained model, to initialize the weights in a new model for the target task. Transfer learning can help to decrease the training time and achieve lower generalization error.

    2. Knee Injury Detection Using Deep Learning on MRI Studies

    Figure 2 shows an increasing trend in adopting ML-based studies in this application area, with most of the papers being published from 2017 onwards (whilst the first ML-based paper on the field was published in 2013). Medical imaging, and specifically MRI, has to be seen as one of the most instructive assets in the field of knee injury diagnosis. The proliferation of MRI data has facilitated the effective training of ML and DL networks towards the development of: (i) novel methodologies that could enhance the medical experts’ domain knowledge and understanding of MRI; and (ii) new, data-driven tools that could enable a more reliable, fast, and fully automated detection of knee injuries. The main characteristics of the proposed MRI-based learning algorithms and pipelines were identified along with the data sources investigated (Table 2).
    Figure 2. Temporal evolution chart depicting the number of ML papers per category published each year since 2013.
    Table 2. Results of studies.

    No.

    Author

    Year

    AI Model Used

    Pretrained CNN

    MRI (T)

    Localization Technique

    Validation

    Performance (Accuracy/AUC)

    Application Domain

    1

    Awan et al. [46]

    2021

    CNN

    ResNet-14

    1.5 T

    They applied normal approach to localize based upon region of interest (ROI)

    5-fold cross-validation

    92%/(healthy tear = 0.98, partial tear = 0.97 and fully ruptured tear = 0.99)

    ACL tear

    2

    Jeon et al. [47]

    2021

    3D CNN

    VGGNet, AlexNet, and SqueezeNet

    3 T & 1.5 T

    Custom localization technique

    5-fold cross-validation

    N/A/0.983 and 0.980 on the

    Chiba and Stanford knee datasets, respectively

    ACL tear

    3

    Rizk et al. [48]

    2021

    3D CNN

    CNN-based localization model

    1 T (54%)–1.5 T (9.7%)–3 T (36.3%)

    Custom localization technique

    ten-fold cross validation

    Meidal = N/A/0.93, Lateral = N/A/0.84

    Meniscus tear

    4

    Dai et al. [49]

    2021

    TransMed

    N/A

    3 T & 1.5 T

    N/A

    120 exams

    ACL tear = 94.9%/0.98, Abnormality = 91.8%/0.976, Meniscus tear = 85.3%/0.95

    ACL tear—Meniscus tear—Abnormalities

    5

    Astuto et al. [50]

    2021

    3D CNN

    N/A

    3 T

    V-Net

    Hold out (15% of sample)

    N/A/from 0.83 to 0.93

    ACL tear—Meniscus tear—Cartilage Lession

    6

    Fritz et al. [15]

    2020

    DCNN

    N/A

    1.5 T (64%)–3 T (36%)

    To visually localize the tear, the software computes the class activation map (CAM) of the last convolution layer in the CNN and maps it to an axial knee image

    Hold out (10% of sample)

    Medial = (86%/0.88), Lateral = (84%/0.78), Overall = (N/A/0.96)

    Meniscus tear

    7

    Namiri et al. [51]

    2020

    CNN

    N/A

    3 T

    three-dimensional V-Net

    Hold out (10% of sample)

    3D-model = (89%/sensitivity of 89% and specificity of 88%), 2D-model = (92%/sensitivity of 93% and specificity of 90%)

    ACL tear

    8

    Zhang et al. [6]

    2020

    CNN

    3D DenseNet, VGG16, ResNet

    1.5 T (74%)–3 T (26%)

    -

    Hold out (20% of sample)

    Custom = (95.7%/0.96), ResNet = (NA/0.95), VGG16 = (NA/0.86)

    ACL tear

    9

    Germann et al. [24]

    2020

    DCNN

    N/A

    1.5 T–3 T

    They cropped manually

    Out of the 5802 MRI studies, 4802 were used for training, 500 for validation, and 500 for initial testing

    N/A/0.94

    ACL tear

    10

    Azcona et al. [52]

    2020

    CNN

    MRNet, ResNet18, Resnet50 and ResNet152, ImageNet

    3 T (56.6%)–1.5 T (43.4%)

    -

    N/A

    NA/0.96–N/A/0.91–N/A/0.94

    ACL tear—Meniscus tear—Abnormalities

    11

    Chang et al. [8]

    2019

    CNN

    ResNet

    1.5 T–3 T

    The object localization CNN was implemented as a fully convolutional network based on U-net architecture

    5-fold-cross-validation

    96.7%/0.97

    ACL tear

    12

    Liu et al. [53]

    2019

    CNN

    LeNet-5, DenseNet, VGG16, AlexNet

    N/A

    They used object detection technique YOLO

    50 subjects test set (14% of the sample)

    N/A/0.98

    ACL tear

    13

    Couteaux et al. [54]

    2019

    CNN

    ResNet-101, ConvNet, R-CNN

    N/A

    To localize both menisci and identify tears in each meniscus, they used the Mask R-CNN framework

    54 cases and the model with the highest validation accuracy was selected

    N/A/0.90

    Meniscus tear

    14

    Pedoia et al. [55]

    2019

    2D U-Net, CNN

    N/A

    3 T

    -

    Hold out (20% of sample)

    Sensitivity of 89.81% and specificity of 81.98%

    Meniscus tear

    15

    Roblot et al. [56]

    2019

    CNN

    AlexNet, MRNet

    N/A

    They used object detection technique Fast RCNN & Faster RCNN

    The algorithm was thus used on a test dataset composed of 700 images for external validation

    72.5%/0.85

    Meniscus tear

    16

    Nicholas Bien et al. [27]

    2018

    CNN

    AlexNET, MRNet

    3 T (56.6%)–1.5 T (43.4%)

    -

    120 exams

    86.7%/0.97–72.5%/0.85–N/A/0.94

    ACL tear—Meniscus tear—Abnormalities

    17

    Liu et al. [57]

    2018

    CNN

    VGG16

    3 T

    -

    fellowship trained musculoskeletal radiologist (R.K., with 15 years of clinical experience)

    N/A/0.92

    Cartilage lesion

    18

    Stajduhar et al. [58]

    2017

    HOG + linSVM, HOG + RF, GIST + rbfSVM, GIST + RF

    N/A

    1.5 T

    Manual extraction of a rectangular ROI

    10-fold cross validation

    (Injury detection problem, complete rupture) = (N/A/0.89, N/A/0.94), (N/A/0.88, N/A/0.94), (N/A/0.889, N/A/0.91), (N/A/0.88, N/A/0.90) respectively with the models

    ACL tear

    19

    Mazlan et al. [59]

    2017

    SVM

    N/A

    N/A

    They use cropping technique

    Hold out (10% of sample)

    100%/N/A

    ACL tear

    20

    Zarandi et al. [60]

    2016

    IT2FCM, PNN

    N/A

    N/A

    -

    Hold out (20% of sample)

    0 and 1 mode: 90%/N/A

    Binary mode: 78%/N/A

    Meniscus tear

    21

    Fu et al. [61]

    2013

    SVM

    N/A

    N/A

    Active Contours without Edges method. This method combines Active Contours with Level Sets and is called ACLS

    5-Fold cross validation

    SVM model: N/A/0.73

    SFFS + SVM: N/A/0.91

    Meniscus tear

    22

    Abdullah et al. [62]

    2013

    BP ANN, K-NN

    N/A

    N/A

    -

    5-fold and 6-fold

    BP ANN: 94.44%/N/A

    k-NN: 87.83%/N/A

    ACL tear

    Although there is no clear acceptance of a “gold-standard” methodological pipeline for diagnosing knee abnormalities using MRI data, it was observed that a number of processing steps were commonly employed in the majority of the reported studies. Figure 3 visualizes a DL pipeline that was adopted by most of the papers, including a pre-processing step, localization (optionally) by identifying regions of interest, and, finally, a CNN-based classification step. Data augmentation was employed by a significant number of papers in the detection of ACL injuries [6][27][46][47][49][50][51][52], in papers where meniscus injuries were investigated [27][49][50][52][55][56], and, finally, in studies focusing on cartilage lesion abnormalities [27][52]. In particular, the available MRI images were modified (via a number of image transformations such as random rotations, shifting, flipping, and the addition of noise) to expand the training dataset, and thus help to improve the performance and ability of the employed DL models to generalize. Localization was employed in papers from all three subcategories: (i) ACL studies [6][8][24][46][47][50][51][53][58]; (ii) meniscus injuries detection studies [15][48][50][54][55][56][60][61]; and (iii) for diagnosing lesion abnormalities [57]. Segmentation or objection detection algorithms were applied in the aforementioned studies to extract areas of interest, enabling the application of CNN-based models on focused and more relevant parts of the initially available images. Given that the region of interest (ROI) may appear in slightly different positions within an image and may have different aspect ratios or sizes, identifying ROIs with an automatic manner has been proven to be a crucial processing step.
    Figure 3. A typical DL pipeline for ACL detection.
    CNN-inspired networks were identified as the dominant approach in the task of extracting informative features from either ROIs or entire MRIs and finally classifying them as normal (healthy) or abnormal (indicating either partial or complete tears). Transfer learning was preferred in most of the cases, allowing the training of big and powerful deep architectures, even if the amount of available data was limited. As networks require a lot of information to be trained from scratch, this technique essentially ‘steals’ knowledge from already pre-trained large networks. Specifically, ResNet variants were used in five papers [6][8][46][52][54] here, whereas VGG [34], AlexNet [33], and MRNet [27] were used three times [6][27][52][53][56][57]. Other pre-trained networks that were used at least once in this survey are: DenseNet [32], Le-Net [63], ImageNet [33], and R-CNN [41]. In five [58][59][60][61][62] out of the 22 studies of the present survey, more traditional ML pipelines were applied, including a separate feature engineering step (where features were manually extracted from images). SVM classification was the preferred classifier in most of the cases.
    Despite the excellent capability of CNNs to come up with valuable image representations, these models lack the capacity for capturing long-range relationships. To deal with this limitation, recent research studies [44][64] have proposed employing Transformer-based architectures for various image recognition tasks. The Transformer [65] is a neural network architecture that relies on global self-attention mechanisms, and it was initially designed for sequence-to-sequence prediction. Papers that used this architectural paradigm have indeed achieved state-of the-art results [66][67] in many natural language processing (NLP) tasks. Dai et al. [49] were the first to employ a Transformer-based architecture for the MRI-based knee injury detection task. In particular, their hybrid (Transformer and CNN) model was used to extract features that pick up the long-range dependencies between MRI and other modalities.

    3. Conclusion

    Notwithstanding the huge potential of AI to improve the medical domain, the DL-based methods have yet to achieve significant deployment in clinical environments. This mainly ensues as a result of: (i) the intrinsic black-box nature of the DL algorithms; and (ii) the high computational cost. Explainable AI aims at building trust in the AI algorithms by providing medical experts with a diagnostic rationale behind the AI decision processes. The goal of the lightweight DL field is to develop models that have shallower architecture and are also faster and more data-efficient, while retaining the high-performance standards. Jeon et al. [47] were the first to get to grips with the clinical deployment of the MRI-based knee injury diagnosis. To this end, they proposed to use post-inference visualisation tools (such as CAM and Grad-CAM), and they also incorporated attention modules, Gaussian positional encoding, squeeze modules, and fewer convolutional filters.

    References

    1. Musahl, V.; Karlsson, J. Anterior cruciate ligament tear. N. Engl. J. Med. 2019, 380, 2341–2348.
    2. Ahmed, I.; Bowes, M.; Hutchinson, C.E.; Parsons, N.; Staniszewska, S.; Price, A.J.; Metcalfe, A. Protocol: Meniscal tear outcome Study (METRO Study): A study protocol for a multicentre prospective cohort study exploring the factors which affect outcomes in patients with a meniscal tear. BMJ Open 2020, 10, e038681.
    3. Darrow, C.J.; Collins, C.L.; Yard, E.E.; Comstock, R.D. Epidemiology of severe injuries among United States high school athletes: 2005–2007. Am. J. Sports Med. 2009, 37, 1798–1805.
    4. Gage, B.E.; McIlvain, N.M.; Collins, C.L.; Fields, S.K.; Dawn Comstock, R. Epidemiology of 6.6 million knee injuries presenting to United States emergency departments from 1999 through 2008. Acad. Emerg. Med. 2012, 19, 378–385.
    5. Merkely, G.; Ackermann, J.; Lattermann, C. Articular cartilage defects: Incidence, diagnosis, and natural history. Oper. Tech. Sports Med. 2018, 26, 156–161.
    6. Zhang, L.; Li, M.; Zhou, Y.; Lu, G.; Zhou, Q. Deep Learning Approach for Anterior Cruciate Ligament Lesion Detection: Evaluation of Diagnostic Performance Using Arthroscopy as the Reference Standard. J. Magn. Reson. Imaging 2020, 52, 1745–1752.
    7. Kaeding, C.C.; Léger-St-Jean, B.; Magnussen, R.A. Epidemiology and diagnosis of anterior cruciate ligament injuries. Clin. Sports Med. 2017, 36, 1–8.
    8. Chang, P.D.; Wong, T.T.; Rasiej, M.J. Deep Learning for Detection of Complete Anterior Cruciate Ligament Tear. J. Digit. Imaging 2019, 32, 980–986.
    9. Logerstedt, D.S.; Snyder-Mackler, L.; Ritter, R.C.; Axe, M.J.; Godges, J.; Altman, R.D.; Briggs, M.; Chu, C.; Delitto, A.; Ferland, A. Knee pain and mobility impairments: Meniscal and articular cartilage lesions: Clinical practice guidelines linked to the international classification of functioning, disability, and health from the orthopaedic section of the American Physical Therapy Association. J. Orthop. Sports Phys. Ther. 2010, 40, A1–A35.
    10. Mather, R.C., III; Koenig, L.; Kocher, M.S.; Dall, T.M.; Gallo, P.; Scott, D.J.; Bach Jr, B.R.; Spindler, K.P.; Group, M.K. Societal and economic impact of anterior cruciate ligament tears. J. Bone Jt. Surg. Am. Vol. 2013, 95, 1751.
    11. Cameron, K.L.; Thompson, B.S.; Peck, K.Y.; Owens, B.D.; Marshall, S.W.; Svoboda, S.J. Normative values for the KOOS and WOMAC in a young athletic population: History of knee ligament injury is associated with lower scores. Am. J. Sports Med. 2013, 41, 582–589.
    12. Huffman, G.R.; Park, J.; Roser-Jones, C.; Sennett, B.J.; Yagnik, G.; Webner, D. Normative SF-36 values in competing NCAA intercollegiate athletes differ from values in the general population. JBJS 2008, 90, 471–476.
    13. Lam, K.C.; Thomas, S.S.; Valier, A.R.S.; McLeod, T.C.V.; Bay, R.C. Previous knee injury and health-related quality of life in collegiate athletes. J. Athl. Train. 2017, 52, 534–540.
    14. Pouly, M.; Koller, T.; Gottfrois, P.; Lionetti, S. Artificial intelligence in image analysis-fundamentals and new developments. Der Hautarzt Z. Fur Dermatol. Venerol. Und Verwandte Geb. 2020, 71, 660–668.
    15. Fritz, B.; Marbach, G.; Civardi, F.; Fucentese, S.F.; Pfirrmann, C.W. Deep convolutional neural network-based detection of meniscus tears: Comparison with radiologists and surgery as standard of reference. Skelet. Radiol. 2020, 49, 1207–1217.
    16. Garwood, E.R.; Tai, R.; Joshi, G. The Use of Artificial Intelligence in the Evaluation of Knee Pathology. In Seminars in Musculoskeletal Radiology; Thieme Medical Publishers: New York, NY, USA, 2020; pp. 21–29.
    17. Palermi, S.; Massa, B.; Vecchiato, M.; Mazza, F.; De Blasiis, P.; Romano, A.M.; Di Salvatore, M.G.; Della Valle, E.; Tarantino, D.; Ruosi, C.J.J.O.F.M.; et al. Indirect Structural Muscle Injuries of Lower Limb: Rehabilitation and Therapeutic Exercise. J. Funct. Morphol. Kinesiol. 2021, 6, 75.
    18. Sirico, F.; Palermi, S.; Massa, B.; Corrado, B. Tendinopathies of the hip and pelvis in athletes: A narrative review. J. Hum. Sports Exerc. 2020, 15, S748–S762.
    19. Hetsroni, I.; Lyman, S.; Do, H.; Mann, G.; Marx, R. Symptomatic pulmonary embolism after outpatient arthroscopic procedures of the knee: The incidence and risk factors in 418 323 arthroscopies. J. Bone Jt. Surg. Br. Vol. 2011, 93, 47–51.
    20. Alanazi, H.O.; Abdullah, A.H.; Qureshi, K.N. A critical review for developing accurate and dynamic predictive models using machine learning methods in medicine and health care. J. Med. Syst. 2017, 41, 69.
    21. Prickett, W.D.; Ward, S.I.; Matava, M.J. Magnetic resonance imaging of the knee. Sports Med. 2001, 31, 997–1019.
    22. Krampla, W.; Roesel, M.; Svoboda, K.; Nachbagauer, A.; Gschwantler, M.; Hruby, W. MRI of the knee: How do field strength and radiologist’s experience influence diagnostic accuracy and interobserver correlation in assessing chondral and meniscal lesions and the integrity of the anterior cruciate ligament? Eur. Radiol. 2009, 19, 1519–1528.
    23. Mohankumar, R.; White, L.M.; Naraghi, A. Pitfalls and pearls in MRI of the knee. Am. J. Roentgenol. 2014, 203, 516–530.
    24. Germann, C.; Marbach, G.; Civardi, F.; Fucentese, S.F.; Fritz, J.; Sutter, R.; Pfirrmann, C.W.; Fritz, B. Deep Convolutional Neural Network–Based Diagnosis of Anterior Cruciate Ligament Tears: Performance Comparison of Homogenous Versus Heterogeneous Knee MRI Cohorts With Different Pulse Sequence Protocols and 1.5-T and 3-T Magnetic Field Strengths. Investig. Radiol. 2020, 55, 499.
    25. Gyftopoulos, S.; Lin, D.; Knoll, F.; Doshi, A.M.; Rodrigues, T.C.; Recht, M.P. Artificial intelligence in musculoskeletal imaging: Current status and future directions. Am. J. Roentgenol. 2019, 213, 506–513.
    26. Shen, D.; Wu, G.; Suk, H.-I. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248.
    27. Bien, N.; Rajpurkar, P.; Ball, R.L.; Irvin, J.; Park, A.; Jones, E.; Bereket, M.; Patel, B.N.; Yeom, K.W.; Shpanskaya, K. Deep-learning-assisted diagnosis for knee magnetic resonance imaging: Development and retrospective validation of MRNet. PLoS Med. 2018, 15, e1002699.
    28. Langerhuizen, D.W.; Janssen, S.J.; Mallee, W.H.; Van Den Bekerom, M.P.; Ring, D.; Kerkhoffs, G.M.; Jaarsma, R.L.; Doornberg, J.N. What are the applications and limitations of artificial intelligence for fracture detection and classification in orthopaedic trauma imaging? A systematic review. Clin. Orthop. Relat. Res. 2019, 477, 2482.
    29. Kunze, K.N.; Rossi, D.M.; White, G.M.; Karhade, A.V.; Deng, J.; Williams, B.T.; Chahla, J. Diagnostic Performance of Artificial Intelligence for Detection of Anterior Cruciate Ligament and Meniscus Tears: A Systematic Review. Arthrosc. J. Arthrosc. Relat. Surg. 2020, 37, 771–781.
    30. Hellerstein, J.M.; Naughton, J.F.; Pfeffer, A. Generalized Search Trees for Database Systems. In Proceedings of the 21st VLDB Conference, Zurich, Switzerland, 11–15 September 1995.
    31. Liu, F. Susan: Segment unannotated image structure using adversarial network. Magn. Reson. Med. 2019, 81, 3330–3345.
    32. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708.
    33. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Processing Syst. 2012, 25, 1097–1105.
    34. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556.
    35. Freeman, W.T.; Roth, M. Orientation histograms for hand gesture recognition. In Proceedings of the International Workshop on Automatic Face and Gesture Recognition, Zurich, Switzerland, 26–28 June 1995; pp. 296–301.
    36. Sebastian, V.B.; Unnikrishnan, A.; Balakrishnan, K. Gray level co-occurrence matrices: Generalisation and some new features. arXiv 2012, arXiv:1205.4831.
    37. Peterson, L.E. K-nearest neighbor. Scholarpedia 2009, 4, 1883.
    38. Hearst, M.A.; Dumais, S.T.; Osuna, E.; Platt, J.; Scholkopf, B. Support vector machines. IEEE Intell. Syst. Appl. 1998, 13, 18–28.
    39. Hassoun, M.H. Fundamentals of Artificial Neural Networks; MIT Press: Cambridge, MA, USA, 1995.
    40. O’Shea, K.; Nash, R. An introduction to convolutional neural networks. arXiv 2015, arXiv:1511.08458.
    41. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587.
    42. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778.
    43. Kamnitsas, K.; Ledig, C.; Newcombe, V.F.; Simpson, J.P.; Kane, A.D.; Menon, D.K.; Rueckert, D.; Glocker, B. Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 2017, 36, 61–78.
    44. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.J.A.P.A. An image is worth 16×16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929.
    45. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980.
    46. Awan, M.J.; Rahim, M.S.M.; Salim, N.; Mohammed, M.A.; Garcia-Zapirain, B.; Abdulkareem, K.H. Efficient detection of knee anterior cruciate ligament from magnetic resonance imaging using deep learning approach. Diagnostics 2021, 11, 105.
    47. Jeon, Y.S.; Yoshino, K.; Hagiwara, S.; Watanabe, A.; Quek, S.T.; Yoshioka, H.; Feng, M.J.I.J.O.B.; Informatics, H. Interpretable and lightweight 3-D deep learning model for automated ACL diagnosis. IEEE J. Biomed. Health Inform. 2021, 25, 2388–2397.
    48. Rizk, B.; Brat, H.; Zille, P.; Guillin, R.; Pouchy, C.; Adam, C.; Ardon, R.; d’Assignies, G.J.P.M. Meniscal lesion detection and characterization in adult knee MRI: A deep learning model approach with external validation. Phys. Med. 2021, 83, 64–71.
    49. Dai, Y.; Gao, Y.; Liu, F.J.D. Transmed: Transformers advance multi-modal medical image classification. Diagnostics 2021, 11, 1384.
    50. Astuto, B.; Flament, I.K.; Namiri, N.; Shah, R.; Bharadwaj, U.; M. Link, T.; D. Bucknor, M.; Pedoia, V.; Majumdar, S.J.R.A.I. Automatic Deep Learning–assisted Detection and Grading of Abnormalities in Knee MRI Studies. Radiol. Artif. Intell. 2021, 3, e200165.
    51. Namiri, N.K.; Flament, I.; Astuto, B.; Shah, R.; Tibrewala, R.; Caliva, F.; Link, T.M.; Pedoia, V.; Majumdar, S. Deep Learning for Hierarchical Severity Staging of Anterior Cruciate Ligament Injuries from MRI. Radiol. Artif. Intell. 2020, 2, e190207.
    52. Azcona, D.; McGuinness, K.; Smeaton, A.F. A Comparative Study of Existing and New Deep Learning Methods for Detecting Knee Injuries using the MRNet Dataset. In Proceedings of the 2020 International Conference on Intelligent Data Science Technologies and Applications (IDSTA), Kuala Lumpur, Malaysia, 18–20 September 2020; pp. 149–155.
    53. Liu, F.; Guan, B.; Zhou, Z.; Samsonov, A.; Rosas, H.; Lian, K.; Sharma, R.; Kanarek, A.; Kim, J.; Guermazi, A. Fully automated diagnosis of anterior cruciate ligament tears on knee MR images by using deep learning. Radiol. Artif. Intell. 2019, 1, 180091.
    54. Couteaux, V.; Si-Mohamed, S.; Nempont, O.; Lefevre, T.; Popoff, A.; Pizaine, G.; Villain, N.; Bloch, I.; Cotten, A.; Boussel, L. Automatic knee meniscus tear detection and orientation classification with Mask-RCNN. Diagn. Interv. Imaging 2019, 100, 235–242.
    55. Pedoia, V.; Norman, B.; Mehany, S.N.; Bucknor, M.D.; Link, T.M.; Majumdar, S. 3D convolutional neural networks for detection and severity staging of meniscus and PFJ cartilage morphological degenerative changes in osteoarthritis and anterior cruciate ligament subjects. J. Magn. Reson. Imaging 2019, 49, 400–410.
    56. Roblot, V.; Giret, Y.; Antoun, M.B.; Morillot, C.; Chassin, X.; Cotten, A.; Zerbib, J.; Fournier, L. Artificial intelligence to diagnose meniscus tears on MRI. Diagn. Interv. Imaging 2019, 100, 243–249.
    57. Liu, F.; Zhou, Z.; Samsonov, A.; Blankenbaker, D.; Larison, W.; Kanarek, A.; Lian, K.; Kambhampati, S.; Kijowski, R. Deep learning approach for evaluating knee MR images: Achieving high diagnostic performance for cartilage lesion detection. Radiology 2018, 289, 160–169.
    58. Štajduhar, I.; Mamula, M.; Miletić, D.; Ünal, G. Semi-automated detection of anterior cruciate ligament injury from MRI. Comput. Methods Programs Biomed. 2017, 140, 151–164.
    59. Mazlan, S.S.; Ayob, M.; Bakti, Z.K. Anterior cruciate ligament (ACL) injury classification system using support vector machine (SVM). In Proceedings of the 2017 International Conference on Engineering Technology and Technopreneurship (ICE2T), Kuala Lumpur, Malaysia, 18–20 September 2017; pp. 1–5.
    60. Zarandi, M.F.; Khadangi, A.; Karimi, F.; Turksen, I. A computer-aided type-II fuzzy image processing for diagnosis of meniscus tear. J. Digit. Imaging 2016, 29, 677–695.
    61. Fu, J.-C.; Lin, C.-C.; Wang, C.-N.; Ou, Y.-K. Computer-aided diagnosis for knee meniscus tears in magnetic resonance imaging. J. Ind. Prod. Eng. 2013, 30, 67–77.
    62. Abdullah, A.A.; Azz-Zahra Md Som, N.S.F. Design of an Intelligent Diagnostic System for Detection of Knee Injuries. Appl. Mech. Mater. 2013, 399, 219–224.
    63. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324.
    64. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P.H. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 22–25 November 2021; pp. 6881–6890.
    65. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008.
    66. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K.J.A.P.A. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805.
    67. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.J.A.P.A. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901.
    More
    Information
    Contributors MDPI registered users' name will be linked to their SciProfiles pages. To register with us, please refer to https://encyclopedia.pub/register : , , , , ,
    View Times: 358
    Revisions: 2 times (View History)
    Update Date: 20 May 2022
    Table of Contents
      1000/1000

      Confirm

      Are you sure to Delete?

      Video Upload Options

      Do you have a full video?
      Cite
      If you have any further questions, please contact Encyclopedia Editorial Office.
      Moustakidis, S.; Siouras, A.; Chalatsis, G.; Liampas, I.; Hantes, M.; Tsaopoulos, D. Knee Injury Detection Using Deep Learning. Encyclopedia. Available online: https://encyclopedia.pub/entry/23157 (accessed on 07 June 2023).
      Moustakidis S, Siouras A, Chalatsis G, Liampas I, Hantes M, Tsaopoulos D. Knee Injury Detection Using Deep Learning. Encyclopedia. Available at: https://encyclopedia.pub/entry/23157. Accessed June 07, 2023.
      Moustakidis, Serafeim, Athanasios Siouras, Georgios Chalatsis, Ioannis Liampas, Michael Hantes, Dimitrios Tsaopoulos. "Knee Injury Detection Using Deep Learning" Encyclopedia, https://encyclopedia.pub/entry/23157 (accessed June 07, 2023).
      Moustakidis, S., Siouras, A., Chalatsis, G., Liampas, I., Hantes, M., & Tsaopoulos, D. (2022, May 20). Knee Injury Detection Using Deep Learning. In Encyclopedia. https://encyclopedia.pub/entry/23157
      Moustakidis, Serafeim, et al. "Knee Injury Detection Using Deep Learning." Encyclopedia. Web. 20 May, 2022.
      Top