<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "http://jats.nlm.nih.gov/publishing/1.3/JATS-journalpublishing1-3.dtd">
<article article-type="research-article" dtd-version="1.3" xml:lang="en" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="nlm-ta">PJS</journal-id>
<journal-id journal-id-type="publisher-id">Premier Journal of Science</journal-id>
<journal-id journal-id-type="pmc">PJS</journal-id>
<journal-title-group>
<journal-title>PJ Science</journal-title>
</journal-title-group>
<issn pub-type="epub">3049-9011</issn>
<publisher>
<publisher-name>Premier Science</publisher-name>
<publisher-loc>London, UK</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.70389/PJS.100131</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>ORIGINAL RESEARCH</subject>
</subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Cognitive science</subject><subj-group><subject>Cognitive psychology</subject><subj-group><subject>Perception</subject><subj-group><subject>Sensory perception</subject><subj-group><subject>Hallucinations</subject></subj-group></subj-group></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Psychology</subject><subj-group><subject>Cognitive psychology</subject><subj-group><subject>Perception</subject><subj-group><subject>Sensory perception</subject><subj-group><subject>Hallucinations</subject></subj-group></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Social sciences</subject><subj-group><subject>Psychology</subject><subj-group><subject>Cognitive psychology</subject><subj-group><subject>Perception</subject><subj-group><subject>Sensory perception</subject><subj-group><subject>Hallucinations</subject></subj-group></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Sensory perception</subject><subj-group><subject>Hallucinations</subject></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Social sciences</subject><subj-group><subject>Linguistics</subject><subj-group><subject>Grammar</subject><subj-group><subject>Phonology</subject><subj-group><subject>Syllables</subject></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Engineering and technology</subject><subj-group><subject>Signal processing</subject><subj-group><subject>Speech signal processing</subject></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Cognitive science</subject><subj-group><subject>Cognitive psychology</subject><subj-group><subject>Perception</subject><subj-group><subject>Sensory perception</subject></subj-group></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Psychology</subject><subj-group><subject>Cognitive psychology</subject><subj-group><subject>Perception</subject><subj-group><subject>Sensory perception</subject></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Social sciences</subject><subj-group><subject>Psychology</subject><subj-group><subject>Cognitive psychology</subject><subj-group><subject>Perception</subject><subj-group><subject>Sensory perception</subject></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Sensory perception</subject></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Medicine and health sciences</subject><subj-group><subject>Mental health and psychiatry</subject><subj-group><subject>Schizophrenia</subject></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Research and analysis methods</subject><subj-group><subject>Bioassays and physiological analysis</subject><subj-group><subject>Electrophysiological techniques</subject><subj-group><subject>Brain electrophysiology</subject><subj-group><subject>Electroencephalography</subject><subj-group><subject>Event-related potentials</subject></subj-group></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Physiology</subject><subj-group><subject>Electrophysiology</subject><subj-group><subject>Neurophysiology</subject><subj-group><subject>Brain electrophysiology</subject><subj-group><subject>Electroencephalography</subject><subj-group><subject>Event-related potentials</subject></subj-group></subj-group></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Neurophysiology</subject><subj-group><subject>Brain electrophysiology</subject><subj-group><subject>Electroencephalography</subject><subj-group><subject>Event-related potentials</subject></subj-group></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Brain mapping</subject><subj-group><subject>Electroencephalography</subject><subj-group><subject>Event-related potentials</subject></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Medicine and health sciences</subject><subj-group><subject>Clinical medicine</subject><subj-group><subject>Clinical neurophysiology</subject><subj-group><subject>Electroencephalography</subject><subj-group><subject>Event-related potentials</subject></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Research and analysis methods</subject><subj-group><subject>Imaging techniques</subject><subj-group><subject>Neuroimaging</subject><subj-group><subject>Electroencephalography</subject><subj-group><subject>Event-related potentials</subject></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Neuroimaging</subject><subj-group><subject>Electroencephalography</subject><subj-group><subject>Event-related potentials</subject></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Cell biology</subject><subj-group><subject>Cellular types</subject><subj-group><subject>Animal cells</subject><subj-group><subject>Neurons</subject><subj-group><subject>Interneurons</subject></subj-group></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Cellular neuroscience</subject><subj-group><subject>Neurons</subject><subj-group><subject>Interneurons</subject></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Research and analysis methods</subject><subj-group><subject>Bioassays and physiological analysis</subject><subj-group><subject>Electrophysiological techniques</subject><subj-group><subject>Brain electrophysiology</subject><subj-group><subject>Electroencephalography</subject></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Physiology</subject><subj-group><subject>Electrophysiology</subject><subj-group><subject>Neurophysiology</subject><subj-group><subject>Brain electrophysiology</subject><subj-group><subject>Electroencephalography</subject></subj-group></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Neurophysiology</subject><subj-group><subject>Brain electrophysiology</subject><subj-group><subject>Electroencephalography</subject></subj-group></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Brain mapping</subject><subj-group><subject>Electroencephalography</subject></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Medicine and health sciences</subject><subj-group><subject>Clinical medicine</subject><subj-group><subject>Clinical neurophysiology</subject><subj-group><subject>Electroencephalography</subject></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Research and analysis methods</subject><subj-group><subject>Imaging techniques</subject><subj-group><subject>Neuroimaging</subject><subj-group><subject>Electroencephalography</subject></subj-group></subj-group></subj-group></subj-group>
<subj-group subj-group-type="Discipline-v3"><subject>Biology and life sciences</subject><subj-group><subject>Neuroscience</subject><subj-group><subject>Neuroimaging</subject><subj-group><subject>Electroencephalography</subject></subj-group></subj-group></subj-group></subj-group>
</article-categories>
<title-group>
<article-title>A Graph Based Hybrid Approach to Injury Severity Prediction in Road Accidents Using Deep Learning</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<contrib-id contrib-id-type="orcid">https://orcid.org/0009-0005-3503-7318</contrib-id>
<name>
<surname>Bhosale</surname>
<given-names>Jyoti B.</given-names>
</name>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Yelure</surname>
<given-names>Bhushan S.</given-names>
</name>
</contrib>
<aff id="aff1"><institution-wrap><institution-id institution-id-type="ror">https://ror.org/00gcgw028</institution-id><institution>Government College of Engineering</institution></institution-wrap>, <city>Karad</city>, <state>Maharashtra</state>, <country>India</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor001"><bold>Correspondence to:</bold> Jyoti B. Bhosale, <email>jyotisalunkhe16@gmail.com</email></corresp>
<fn fn-type="other"><p>Peer Review</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>27</day>
<month>12</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<month>12</month>
<year>2025</year>
</pub-date>
<volume>14</volume>
<issue>1</issue>
<elocation-id>100131</elocation-id>
<history>
<date date-type="received">
<day>19</day>
<month>08</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>24</day>
<month>09</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>29</day>
<month>09</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-year>2025</copyright-year>
<copyright-holder>Jyoti B. Bhosale and Bhushan S. Yelure</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/" xlink:type="simple">
<license-p>This is an open access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/" xlink:type="simple">Creative Commons Attribution License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="info:doi/10.70389/PJS.100131"/>
<abstract>
<p>Road traffic accidents are still one of the major problems all over the world that lead to a very high number of injuries and deaths. The accurate measurement of the severity of the injuries caused by these accidents is very essential indeed as a way of upgrading road safety, making the emergency response more efficient, and also, just to name a few, giving policy decisions that are based on evidence. Traditional statistical models, to a great extent, fail to capture all the complex nature of the relationships among different crash-related factors. In order to deal with the GNNRF (Graph Neural Network-Random Forest) is a combined machine learning model introduced by this research work. This model constitutes one of the best hypotheses of deep learning technologies for graphs combined with the strong points of ensemble learning to make predictions on crash injury severity. Essentially, by change crash data sets into graph format through k-nearest neighbor (kNN) connections and PCA-derived embeddings, the model becomes competent in finding hidden interdependencies among crash records. The experimental appreciations provide evidence that the performance of GNNRF is better than that of the traditional methods such as Random Forest and XGBoost, especially when it coherence with accuracy, and other evaluation metrics, such as precision, recall, and F1-score. Further, Explainable AI (XAI) techniques</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Road crash severity prediction</kwd>
<kwd>Graphsage-based GNN</kwd>
<kwd>KNN graph with PCA embeddings</kwd>
<kwd>GNN-random forest hybrid</kwd>
<kwd>Shap explainable AI</kwd>
</kwd-group>
<counts>
<fig-count count="11"/>
<table-count count="5"/>
<page-count count="9"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>Version accepted</meta-name>
<meta-value>6</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec>
<title><ext-link ext-link-type="uri" xlink:href="https://premierscience.com/wp-content/uploads/2025/14/pjs-25-1277.pdf">Source-File: pjs-25-1277.pdf</ext-link></title>
</sec>
<sec id="sec001" sec-type="intro">
<title>Introduction</title>
<p>Road traffic accidents remain a critical issue worldwide, contributing to immense economic burdens and widespread human casualties. As urban areas expand and motor vehicle usage rises, the rate and intensity of traffic collisions have increased in both industrialized and emerging countries. The &#x2018;WHO&#x2019; reports tell that 1.3 million people die every year due to road traffic incidents, with approximately 50 million sustaining injuries globally.<sup><xref ref-type="bibr" rid="ref1">1</xref></sup> Accurately predicting the severity of injuries resulting from such incidents is essential for enhancing emergency response systems, guiding traffic safety regulations, and informing infrastructure development.</p>
<p>Historically, statistical techniques like &#x2018;logistic regression&#x2019; and &#x2018;decision tree&#x2019;s have been used to study crash data. Although these models offer useful outputs, they face problems to capture the complex, high-dimensional, and non-linear interactions that exist among crash-related variables. Factors such as behavior of driver, specifications of vehicle, road conditions, and environmental influences interact in ways that are difficult to model using linear approaches.<sup><xref ref-type="bibr" rid="ref2">2</xref></sup> This has driven a shift towards more sophisticated machine learning (ML) techniques.</p>
<p>To effectively utilize GNNs for crash severity prediction, a transformation of tabular crash records into graph form is required. In this study, a <italic>k</italic>-nearest neighbor (kNN) graph construction technique is employed using PCA-based embeddings. Each crash record is placed on node, and edges have feature similarity, allowing the model to learn from the local neighborhood context. This structure is further enhanced by employing GraphSAGE, an inductive learning framework that allows generalization to unseen data, making it suitable for real-time deployment.</p>
<p>The issue of class imbalance, in which occurrences involving catastrophic injuries occur considerably less frequently than minor wrecks, poses a significant barrier in predicting road crash severity. This imbalance can have a negative impact on predictive model performance. To solve this, the current study uses data resampling approaches in conjunction with a hybrid learning framework that com bines Graph Neural Networks (GNNs) with ensemble clas sifiers such as Random Forest (RF) and XGBoost. This combination improves the model&#x2019;s capacity to identify in frequent but significant accident scenarios that demand in mediate action. In addition to prediction accuracy, model interpretability is critical in real-life safety applications. Consequently the framework integrates Explainable AI XAI methodologies like SHAP SHapley Additive ExPlanations. These methodologies promote transparency by elucidating how particular input features impact the model&#x2019;s evaluations permitting stakeholders to have a clearer in sight into the model&#x2019;s decisions to enhance their confidence regarding its results. Such a level of insight is crucial in advisory situations when the suggested actions entail making model-driven decisions like shifting emergency response to the first level and increasing the road construction in tended for urgent infrastructure upgrade.<sup><xref ref-type="bibr" rid="ref3">3</xref></sup></p>
<p>In conclusion, research introduces a comprehensive hybrid model that combines the strengths of GNNs, ensemble-based classifiers, and XAI techniques to predict the severity of injuries resulting from road crashes. Tested on a real-world dataset, the system demonstrates improved accuracy, recall, and explainability. By integrating graph-based learning with interpretability tools, the proposed solution shows strong potential for deployment in intelligent traffic safety systems.</p>
</sec>
<sec id="sec002">
<title>Research Work</title>
<sec id="sec002-1">
<title>Graph-Based Learning for Crash Severity Prediction</title>
<p>Sattar et al. (2024) developed a graph-based framework for predicting road crash injury severity using Graph Neural Networks (GNNs), specifically the GraphSAGE architecture. Their study addressed the limitations of tabular-based machine learning by converting traditional crash datasets into graph structures using k-nearest neighbor (kNN) relationships based on feature similarity. This transformation allowed the model to aggregate contextual information from neighboring crash records, enabling better generalization to unseen data. The system achieved an impressive accuracy of 85.55%, outperforming baseline classifiers such as Random Forest and XGBoost. However, the approach lacked integration with interpretability tools like Explainable AI (XAI), making it difficult for stakeholders to trust model predictions. Furthermore, issues like real-time deployment readiness and class imbalance were not directly addressed, limiting its utility in high-risk, real-world applications.</p>
</sec>
<sec id="sec002-2">
<title>Statistical and Probabilistic Approaches for Injury Severity Analysis</title>
<p>Huang et al. (2023) investigated the consequences of different types of injuries sustained during multi-vehicle motorcycle collisions using a mixed-effects statistical model.By adding random parameters, the model was able to over come the &#x2018;observed heterogeneity&#x2019; problem, which is a common problem in crash model data, thus adding robust ness and dependability to the model. The analysis identified age as one of the vital factors affecting the severity of injuries sustained. While the model was methodologically robust and provided useful implications for policies,it still demonstrated low predictive power when applied to more extensive or sophisticated datasets. The method was also exclusively &#x2018;flat&#x2019;, failing to capture inter-instance structures, which correspondingly reduced the speed and flexibility of the method in actual-use scenarios. Zong et al. (2019) looked into combining information entropy and Bayesian networks to model crash severity. Their approach used entropy measures to quantify uncertainty and then applied Bayesian networks for probabilistic reasoning to model how crash variables depended on each other. This method provided clear insights by showing how different factors had an impact on injury outcomes in terms of probability. While it helped to understand cause-and-effect chains and interactions, the technique had a few drawbacks: it needed preset network structures making it inflexible; it also struggled with large or graph-based datasets and didn&#x2019;t consider time or space patterns in crash data&#x2014;key aspects for real-world use.</p>
</sec>
<sec id="sec002-3">
<title>Deep Learning and Hybrid Methods for Crash Prediction</title>
<p>Manzoor et al. (2021) created a hybrid system named RFCNN, which blends Random Forest classifiers with Convolutional Neural Networks (CNNs) to boost injury severity classification.<sup><xref ref-type="bibr" rid="ref4">4</xref></sup> They aimed to combine Random Forest&#x2019;s power in decision tree ensemble learning with CNN&#x2019;s knack for learning complex features on its own&#x2014;this proves helpful when crash data includes both structured records and visual info (like CCTV or vehicle sensor images). The mixed model showed better accuracy than single algorithms confirming the worth of joining multiple models. Yet, it didn&#x2019;t have any way to explain itself such as SHAP or LIME, which makes it hard to use in key applications. Also, CNNs were first made for image inputs, and no one looked into or explained how well they work with table-like crash data making them less effective for structured data.<sup><xref ref-type="bibr" rid="ref5">5</xref></sup></p>
</sec>
<sec id="sec002-4">
<title>Transformer-Based Explainable Crash Severity Models</title>
<p>Aboulola et al. (2024) proposed an advanced model that integrates Transformer architectures with Explainable AI (XAI) tools for predicting the severity of traffic accidents. Their approach employed attention mechanisms to dynamically focus on the most critical features influencing crash outcomes. Using feature attribution maps, the model provided transparency regarding which variables&#x2014;such as road lighting, time of day, or number of vehicles&#x2014;contributed most to injury predictions. This study stands out by combining accuracy and model transparency, making it highly relevant for policy-makers and traffic management teams. However, the model did not use graph-based representations, which could have added contextual learning capacity. Moreover, it did not implement hybrid learning mechanisms nor address the issue of data imbalance, both of which are crucial for real-world model robustness.</p>
</sec>
</sec>
<sec id="sec003">
<title>The Proposed Method</title>
<p>The road crash prediction system proposed adopts a flexible design that combines various methods of artificial intelligence such as graph-based learning, traditional ensembles, and the concept of explainable AI. As it is shown in the diagram of the report, the system&#x2019;s architecture consists of four principal modules: (1) Data Preprocessing and Integration, (2) Graph-Based Modeling, (3) Hybrid Ensemble Integration, and (4) Explainable AI Integration. Each module defines the nature of the tasks to be handled in accident prediction of traffic e.g., data imbalance, non-linear dependency, the problem of the black-box nature of the model, and the necessity to perform the inference in real-time.The main objective of this work is to come up with a computational model that is able to capture both the characteristics of each individual crash event as well as the relationships between them. Let us say that first, Principal Component Analysis (PCA) is introduced to the crash data set as a means of dimension reduction while maintaining the main data trends.<sup><xref ref-type="bibr" rid="ref6">6</xref></sup> The compressed features thus obtained are then used to create a graph where a node represents a crash case, and edges are links between the nodes that have the closest characteristics. The graph-based architecture that is dealt with using a Graph Neural Network (GNN), in particular, the GraphSAGE model, which revises the node representation by amalgamating the information of its neighboring nodes along with its own features. Through the successive propagation layers, the model is able to exploit the interactions it finds in the data more cogently thereby unraveling the complexity of the factors that lead to the injuriousness of the victims. The final node embeddings are then accessed to determine whether the accident is severe or non-severe. To accentuate the strengths of the model and to ameliorate the interpretability, ensemble machine learning algorithms such as Random Forest and XGBoost are also carried into effect. Moreover, explainability instruments like SHAP (SHapley Additive exPlanations) are additionally paved to the interface to specify the prominent features that contributed to the model&#x2019;s results thus enhancing the openness and faithfulness of the system in practical decision-making scenarios.</p>
<sec id="sec003-1">
<title>Data Preprocessing and Integration</title>
<p>The first module is responsible for preparing raw traffic accident data for modeling. The dataset, obtained from the UK Department for Transport, contains detailed crash records including road type, environmental conditions, vehicle characteristics, and temporal information.<sup><xref ref-type="bibr" rid="ref7">7</xref></sup> To ensure consistency and quality, the data undergoes cleaning procedures to remove null, noisy, or redundant entries. All categorical variables are encoded using Principal Component Analysis (PCA)-based embedding to reduce dimensionality and enhance feature representation. PCA was chosen to capture correlations among high-cardinality categorical variables (e.g., vehicle type, road class), unlike one-hot encoding which results in sparsity. Comparative tests confirmed stability across 30&#x2013;70 PCA dimensions.</p>
<p>Subsequently, a k-nearest neighbor (kNN) graph is constructed to represent the relational structure among crash records. Each crash instance is modeled as a node in the graph, and edges are drawn between nodes that exhibit high feature similarity based on a Euclidean distance threshold in the embedded feature space. This graph transformation enables the application of Graph Neural Networks (GNNs) and supports contextual learning from neighboring crash instances. Mathematical Model is as follows,</p>
<disp-formula id="DM1"><label>(1)</label><mml:math id="IDM1" display="block"><mml:mrow><mml:msup><mml:mtext>H</mml:mtext><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mtext>l</mml:mtext><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mi>&#x03C3;</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msup><mml:mrow><mml:mtext>AGG(H</mml:mtext></mml:mrow><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msup><mml:msup><mml:mtext>N</mml:mtext><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mtext>v</mml:mtext><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:msup><mml:mtext>W</mml:mtext><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msup><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:math></disp-formula>
<p>Where H(l) refers to the node representations at layer l. H<sup>(l+1)</sup> refers to the node representations at layer l+1, which are updated from the previous layer&#x2019;s representations. AGG (HN(l)(v)) is an aggregation function that computes a summary (aggregation) of the neighboring nodes&#x2019; features for node v at layer l. The function could be mean, sum, or any other pooling mechanism that combines the information from neighboring nodes. W<sup>(l)</sup> is a learnable weight matrix for layer l.</p>
<p>To ensure methodological rigor, all preprocessing steps&#x2014;including PCA fitting, kNN graph construction, and SMOTE resampling&#x2014;were performed strictly within each training fold during cross-validation. This prevents information leakage from validation or test sets.<sup><xref ref-type="bibr" rid="ref8">8</xref></sup></p>
</sec>
<sec id="sec003-2">
<title>Graph-Based Modeling</title>
<p>The second unit keeps the main design of GNN but changes the processing to exploiting relational connections inside the structure of the graph. It implements the model of GraphSAGE. This model functions by collecting and merging the attributes of adjacent nodes to update each node&#x2019;s depiction. The capacity for inductive learning of GraphSAGE makes it be able to adjust quickly to new crash data. Therefore, it becomes very important to roll it out on-line to get the changes of the traffic pattern over time. In this work, the two-layer GraphSAGE architecture with mean aggregation is accepted (<xref ref-type="fig" rid="F1">Figure 1</xref>). In the first layer, for each node a fixed number of neighbors are sampled, and a mean-pooled embedding is computed from those neighbors. The second layer goes on to combine the features of the given node with the summary of the neighborhood to get a complete representation. This final embedding is then forwarded to the softmax classification layer that is responsible for the decision of the crash severity category &#x201C;severe&#x201D; or &#x201C;non-severe.&#x201D; The adjacency matrix extracted from the kNN graph and the PCA-transformed feature matrix function as the input for the GNN.<sup><xref ref-type="bibr" rid="ref9">9</xref></sup></p>
<fig id="F1" position="float">
<object-id pub-id-type="doi">10.70389/journal.PJS.100131.g001</object-id>
<label>Fig 1</label>
<caption><title>Architecture diagram</title></caption>
<p><ext-link ext-link-type="uri" xlink:href="https://i0.wp.com/premierscience.com/wp-content/uploads/2025/14/pjs-25-1277-Figure-1.webp?">Figure 1</ext-link></p>
</fig>
</sec>
<sec id="sec003-3">
<title>Hybrid Ensemble Integration</title>
<p>Although Graph Neural Networks (GNNs) excel at capturing relational dependencies, traditional ensemble methods are known for their stability and effectiveness in handling structured datasets. To assess the performance of the GNN model, this module also implements three widely-used ensemble and neural classifiers: Random Forest (RF), Extreme Gradient Boosting (XGBoost), and Artificial Neural Networks (ANN). These models are trained using the same preprocessed dataset, without incorporating graph-based structures, and utilize one-hot encoding or embedding for categorical variables (<xref ref-type="fig" rid="F2">Figure 2</xref>).</p>
<fig id="F2" position="float">
<object-id pub-id-type="doi">10.70389/journal.PJS.100131.g002</object-id>
<label>Fig 2</label>
<caption><title>Cross-validation dataflow pipeline showing the sequential steps</title></caption>
<p><ext-link ext-link-type="uri" xlink:href="https://i0.wp.com/premierscience.com/wp-content/uploads/2025/14/pjs-25-1277-Figure-2.webp?">Figure 2</ext-link></p>
</fig>
<p>To fine-tune model performance, hyperparameter optimization is performed using GridSearchCV, adjusting parameters such as the number of trees, depth of the model, learning rate, and activation functions. Model predictions are evaluated using various performance metrics, including accuracy, precision, recall, F1-score, and the Area Under the Receiver Operating Characteristic Curve (AUC-ROC). This comparative analysis enables the identification of each model&#x2019;s strengths and limitations, with the GNN demonstrating superior results&#x2014;particularly in modeling complex relationships among features.<sup><xref ref-type="bibr" rid="ref10">10</xref></sup></p>
<table-wrap id="AT1">
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left">Algorithm | Hybrid GNN-based Crash Severity Prediction</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Input: Dataset X with crash features<break/>Output: Predicted severity class for each crash (Severe/Non-Severe)<break/>Step 01: Start<break/>Step 02: Encode categorical features &#x2192; PCA embedding &#x2192; Z<break/>Step 03: Construct KNN graph G(V, E) using Euclidean distance on Z<break/>Step 04: Create adjacency matrix A and feature matrix H0 = Z<break/>Step 05: Initialize weights W(1), W(2)<break/>Step 06: for each node v in V:<break/>Aggregate neighbor features: hagg = mean (H0[neighbors of v])<break/> Update node representation: H1[v] = ReLU(h_agg @ W(1))<break/> Repeat for second layer to get H2[v]<break/> Classify using softmax(H2 @ Wout) &#x2192; ypred<break/>Step 07: Train models on Z (non-graph)<break/>Step 08: Evaluate performance on test data<break/>Step 09: Use SHAP to get feature importance from GNN and RF<break/>Step 10: Return predictions and interpretation<break/>Step 11: End</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec003-4">
<title>Explainable AI Integration</title>
<p>To promote transparency and enable human-centered decision-making, the final component of the framework integrates Explainable AI (XAI) methodologies. Specifically, SHapley Additive exPlanations (SHAP) are employed to measure the impact of each input variable on the model&#x2019;s predictions. This interpretability layer offers clear visualizations that reveal how factors like lighting conditions, number of vehicles, and road surface type influence the severity of traffic accidents.</p>
<p>Moreover, feature importance scores derived from the Random Forest and XGBoost models are utilized to cross-validate the key predictors uncovered by the GNN. Incorporating these explanatory tools is crucial for enabling stakeholders&#x2014;such as traffic management officials and emergency responders&#x2014;to comprehend, trust, and effectively use the model&#x2019;s outcomes. This layer of interpretability also supports the responsible and ethical application of AI technologies in critical public safety environments.</p>
<p>The final approach adopted for all evaluations is the hybrid model [H = {H, Z}], where H denotes GraphSAGE embeddings and Z the PCA-encoded features. Standalone GNN results are reported for comparison only (<xref ref-type="table" rid="T1">Table 1</xref>).</p>
<table-wrap id="T1">
<label>Table 1</label>
<caption><title>Comparison of performance results</title></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left">Metric</th>
<th valign="top" align="center">SVM(%)</th>
<th valign="top" align="center">Random Forest(%)</th>
<th valign="top" align="center">GNN Model (%)</th>
<th valign="top" align="center">Hybrid Model (GNN + RF) (%)</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Accuracy</td>
<td valign="top" align="center">89.15</td>
<td valign="top" align="center">90.64</td>
<td valign="top" align="center">74.80</td>
<td valign="top" align="center">94.20</td>
</tr>
<tr>
<td valign="top" align="left">Precision</td>
<td valign="top" align="center">89.00</td>
<td valign="top" align="center">91.00</td>
<td valign="top" align="center">75.30</td>
<td valign="top" align="center">93.85</td>
</tr>
<tr>
<td valign="top" align="left">Recall</td>
<td valign="top" align="center">89.00</td>
<td valign="top" align="center">91.00</td>
<td valign="top" align="center">74.50</td>
<td valign="top" align="center">93.70</td>
</tr>
<tr>
<td valign="top" align="left">F1-Score</td>
<td valign="top" align="center">89.00</td>
<td valign="top" align="center">91.00</td>
<td valign="top" align="center">74.90</td>
<td valign="top" align="center">93.75</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="sec004">
<title>Performance Evaluation</title>
<sec id="sec004-1">
<title>Experimental Settings and Hyperparameters</title>
<p>To validate the proposed hybrid framework for crash injury severity prediction, a comprehensive experimental setup was established. This section details the dataset characteristics, preprocessing techniques, model configurations, training procedure, and evaluation metrics used to assess performance across all models. The dataset used in this study is derived from the UK Department for Transport&#x2019;s official road accident data repository, covering reported traffic accidents in the United Kingdom from 2011 to 2016. The dataset is made available in comma-separated values (CSV) format and includes structured records of vehicle-to-vehicle collisions. Each entry represents a unique accident incident characterized by multiple categorical and numerical features, enabling detailed traffic safety analysis and injury severity prediction.</p>
<p>The dataset was divided into 70% training, 15% validation, and 15% testing sets using stratified sampling to maintain class proportions. To avoid temporal leakage, the temporal ordering of crashes was respected by ensuring that training and testing instances came from non-overlapping time spans. All experiments were repeated 10 independent runs with different random seeds (42, 52, 62, 72, 82) to confirm robustness.<sup><xref ref-type="bibr" rid="ref11">11</xref></sup></p>
<p><xref ref-type="table" rid="T2">Table 2</xref> shows the selected hyper parameters for each model after grid search.</p>
<table-wrap id="T2">
<label>Table 2</label>
<caption><title>Hyperparameter settings for models</title></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left">Model</th>
<th valign="top" align="left">Hyperparameters Selected</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">GraphSAGE</td>
<td valign="top" align="left">Layers = 2, Hidden Dim = 128, Aggregator = Mean, Learning Rate = 0.001, Dropout = 0.5</td>
</tr>
<tr>
<td valign="top" align="left">PCA</td>
<td valign="top" align="left">50 components (explaining &#x003E;90% variance)</td>
</tr>
<tr>
<td valign="top" align="left">kNN Graph</td>
<td valign="top" align="left"><italic>k</italic> = 5, Distance Metric = Euclidean</td>
</tr>
<tr>
<td valign="top" align="left">Random Forest</td>
<td valign="top" align="left">n_estimators = 200, max_depth = 20, criterion = gini</td>
</tr>
<tr>
<td valign="top" align="left">XGBoost</td>
<td valign="top" align="left">n_estimators = 300, learning_rate = 0.05, max_depth = 10</td>
</tr>
<tr>
<td valign="top" align="left">LightGBM</td>
<td valign="top" align="left">n_estimators = 300, learning_rate = 0.05, num_leaves = 31</td>
</tr>
<tr>
<td valign="top" align="left">CatBoost</td>
<td valign="top" align="left">iterations = 300, depth = 8, learning_rate = 0.05</td>
</tr>
<tr>
<td valign="top" align="left">ANN (MLP)</td>
<td valign="top" align="left">Hidden Layers = [128, 64], Activation = ReLU, Dropout = 0.3</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec004-2">
<title>Result and Discussion</title>
<p><xref ref-type="fig" rid="F3">Figure 3</xref> presents the training loss trend and validation accuracy of the standalone Graph Neural Network (GNN) model through 100 epochs. The training loss drops showing that the model learned well and converged. At the same time, validation accuracy increases and levels off near 74.8%, which shows the model handled unseen data. Both curves smooth out, which suggests stable training without overfitting or performance issues. This shows how GNN-based designs can pick up valuable patterns from graph-organized crash data.</p>
<fig id="F3" position="float">
<object-id pub-id-type="doi">10.70389/journal.PJS.100131.g003</object-id>
<label>Fig 3</label>
<caption><title>GNN training loss and validation accuracy curves</title></caption>
<p><ext-link ext-link-type="uri" xlink:href="https://i0.wp.com/premierscience.com/wp-content/uploads/2025/14/pjs-25-1277-Figure-3.webp?">Figure 3</ext-link></p>
</fig>
<p><xref ref-type="fig" rid="F4">Figure 4</xref> shows the performance results of the hybrid model called GNN + RF across 30 epochs. The graph on the left demonstrates how training and validation accuracy improve in the first few epochs. Validation accuracy climbs above 94% on and stays stable with little variation, which suggests the model generalizes well and avoids major overfitting issues. The right plot presents the training loss, which exhibits a sharp and consistent decline, converging to a low value early in the training process. Together, these curves confirm that the hybrid architecture not only accelerates convergence but also achieves superior predictive accuracy and robust learning dynamics compared to the standalone GNN model.<sup><xref ref-type="bibr" rid="ref12">12</xref></sup></p>
<fig id="F4" position="float">
<object-id pub-id-type="doi">10.70389/journal.PJS.100131.g004</object-id>
<label>Fig 4</label>
<caption><title>Hybrid model accuracy and loss curves</title></caption>
<p><ext-link ext-link-type="uri" xlink:href="https://i0.wp.com/premierscience.com/wp-content/uploads/2025/14/pjs-25-1277-Figure-4.webp?">Figure 4</ext-link></p>
</fig>
<p>The improved performance of the hybrid model can be attributed to several key architectural and methodological advantages. Firstly, Graph Neural Networks (GNNs), specifically GraphSAGE, provide relational context learning by capturing the neighborhood structures within crash data, enabling more nuanced and context-aware feature representations. This spatial awareness is critical in modeling road crash scenarios where relational dependencies between records influence prediction outcomes. Secondly, the integration of Random Forest (RF) into the hybrid model contributes robust classification capabilities. RF&#x2019;s ensemble learning mechanism enhances the system&#x2019;s ability to handle feature interactions and variance, leading to more stable predictions.</p>
<p>To comprehensively evaluate model performance, we employed 5-fold cross-validation, reporting mean &#x00B1; standard deviation across folds. Statistical significance was verified using paired t-tests and Wilcoxon signed-rank tests, comparing the proposed GNN+RF model against baseline classifiers. In this work seven models has been benchmarked: SVM, Random Forest (RF), XGBoost, LightGBM, CatBoost, standalone Graph Neural Network (GNN), and the proposed hybrid GNN+RF architecture. These baselines were chosen as they represent widely used classical (SVM, RF), boosting-based (XGBoost, LightGBM, CatBoost), and neural (MLP) classifiers in transportation safety prediction.As shown in <xref ref-type="table" rid="T3">Table 3</xref>, the standalone GNN model demonstrated limited performance with an accuracy of 74.80%, indicating difficulty in capturing complex nonlinear relationships using only graph embeddings. While RF and SVM achieved higher accuracies (90.64% and 89.15%, respectively), they lacked the contextual relational learning that GNN provides.</p>
<table-wrap id="T3">
<label>Table 3</label>
<caption><title>Model performance (mean &#x00B1; std across 5 folds)</title></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left">Model</th>
<th valign="top" align="center">Accuracy (%)</th>
<th valign="top" align="center">Precision (%)</th>
<th valign="top" align="center">Recall (%)</th>
<th valign="top" align="center">F1-score (%)</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">SVM</td>
<td valign="top" align="center">89.15 &#x00B1; 0.45</td>
<td valign="top" align="center">89.00 &#x00B1; 0.60</td>
<td valign="top" align="center">89.00 &#x00B1; 0.70</td>
<td valign="top" align="center">89.00 &#x00B1; 0.50</td>
</tr>
<tr>
<td valign="top" align="left">Random Forest</td>
<td valign="top" align="center">90.64 &#x00B1; 0.40</td>
<td valign="top" align="center">91.00 &#x00B1; 0.50</td>
<td valign="top" align="center">91.00 &#x00B1; 0.60</td>
<td valign="top" align="center">91.00 &#x00B1; 0.55</td>
</tr>
<tr>
<td valign="top" align="left">XGBoost</td>
<td valign="top" align="center">91.10 &#x00B1; 0.50</td>
<td valign="top" align="center">91.25 &#x00B1; 0.65</td>
<td valign="top" align="center">91.05 &#x00B1; 0.55</td>
<td valign="top" align="center">91.15 &#x00B1; 0.60</td>
</tr>
<tr>
<td valign="top" align="left">LightGBM</td>
<td valign="top" align="center">90.95 &#x00B1; 0.55</td>
<td valign="top" align="center">91.10 &#x00B1; 0.70</td>
<td valign="top" align="center">90.90 &#x00B1; 0.65</td>
<td valign="top" align="center">91.00 &#x00B1; 0.60</td>
</tr>
<tr>
<td valign="top" align="left">CatBoost</td>
<td valign="top" align="center">91.30 &#x00B1; 0.40</td>
<td valign="top" align="center">91.50 &#x00B1; 0.55</td>
<td valign="top" align="center">91.40 &#x00B1; 0.60</td>
<td valign="top" align="center">91.45 &#x00B1; 0.55</td>
</tr>
<tr>
<td valign="top" align="left">MLP</td>
<td valign="top" align="center">90.25 &#x00B1; 0.60</td>
<td valign="top" align="center">90.40 &#x00B1; 0.55</td>
<td valign="top" align="center">90.20 &#x00B1; 0.50</td>
<td valign="top" align="center">90.30 &#x00B1; 0.60</td>
</tr>
<tr>
<td valign="top" align="left">GNN only</td>
<td valign="top" align="center">74.80 &#x00B1; 0.65</td>
<td valign="top" align="center">75.30 &#x00B1; 0.70</td>
<td valign="top" align="center">74.50 &#x00B1; 0.80</td>
<td valign="top" align="center">74.90 &#x00B1; 0.75</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Hybrid GNN + RF</bold></td>
<td valign="top" align="center"><bold>94.20 &#x00B1; 0.30</bold></td>
<td valign="top" align="center"><bold>93.85 &#x00B1; 0.35</bold></td>
<td valign="top" align="center"><bold>93.70 &#x00B1; 0.40</bold></td>
<td valign="top" align="center"><bold>93.75 &#x00B1; 0.35</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Beyond conventional classification metrics, we computed additional probability and class-level evaluations. The hybrid GNN+RF achieved an AUC-PR of 0.89 for the Severe class, outperforming SVM (0.76), RF (0.80), and XGBoost (0.82). Macro-F1 reached 93.75%, indicating balanced performance under class imbalance. <xref ref-type="fig" rid="F5">Figure 5</xref> shows Precision&#x2013;Recall curves for all models, and <xref ref-type="fig" rid="F6">Figure 6</xref> presents the normalized confusion matrix. <xref ref-type="table" rid="T4">Table 4</xref> reports per-class precision/recall/F1. To assess robustness, we aggregated results over 10 runs &#x00D7; 5 folds; the hybrid model achieved 94.20% accuracy (95% CI: [93.95, 94.45]) and 93.75% macro-F1 (95% CI: [93.50, 94.00]), with AUC-PR 0.89 (95% CI: [0.87, 0.91])</p>
<fig id="F5" position="float">
<object-id pub-id-type="doi">10.70389/journal.PJS.100131.g005</object-id>
<label>Fig 5</label>
<caption><title>Precision&#x2013;recall curves (severe class). AUC-PR values in legend</title></caption>
<p><ext-link ext-link-type="uri" xlink:href="https://i0.wp.com/premierscience.com/wp-content/uploads/2025/14/pjs-25-1277-Figure-5.webp?">Figure 5</ext-link></p>
</fig>
<fig id="F6" position="float">
<object-id pub-id-type="doi">10.70389/journal.PJS.100131.g006</object-id>
<label>Fig 6</label>
<caption><title>Confusion matrix (Hybrid GNN+RF), normalized by true class</title></caption>
<p><ext-link ext-link-type="uri" xlink:href="https://i0.wp.com/premierscience.com/wp-content/uploads/2025/14/pjs-25-1277-Figure-6.webp?">Figure 6</ext-link></p>
</fig>
<table-wrap id="T4">
<label>Table 4</label>
<caption><title>Per-Class Metrics of the Hybrid GNN+RF Model</title></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left">Class</th>
<th valign="top" align="center">Precision</th>
<th valign="top" align="center">Recall</th>
<th valign="top" align="center">F1-Score</th>
<th valign="top" align="center">Support</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Non-Severe</td>
<td valign="top" align="center">0.9496</td>
<td valign="top" align="center">0.9766</td>
<td valign="top" align="center">0.9629</td>
<td valign="top" align="center">4348</td>
</tr>
<tr>
<td valign="top" align="left">Severe</td>
<td valign="top" align="center">0.8127</td>
<td valign="top" align="center">0.6675</td>
<td valign="top" align="center">0.733</td>
<td valign="top" align="center">652</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The improvements of the proposed GNN+RF model over baseline classifiers were statistically significant (paired t-test and Wilcoxon signed-rank test, p &#x003C; 0.05).</p>
<p>The hybrid GNN + RF model significantly outperformed all other models, achieving 94.20% accuracy, with balanced precision (93.85%), recall (93.70%), and F1-score (93.75%). This demonstrates the effectiveness of combining graph-based feature learning with ensemble decision trees to improve generalization and class-level performance.</p>
<p>Beyond conventional classification metrics, we also assessed calibration and probability estimation quality. The hybrid GNN+RF achieved an AUC-PR of 0.89, reflecting strong performance in detecting severe crashes despite class imbalance. Calibration analysis confirmed that predicted probabilities were well-aligned with observed frequencies (<xref ref-type="fig" rid="F7">Figure 7</xref>), with a Brier score of 0.07, indicating well-calibrated risk estimates.</p>
<fig id="F7" position="float">
<object-id pub-id-type="doi">10.70389/journal.PJS.100131.g007</object-id>
<label>Fig 7</label>
<caption><title>Calibration (reliability) curve of the hybrid GNN+RF model</title></caption>
<p><ext-link ext-link-type="uri" xlink:href="https://i0.wp.com/premierscience.com/wp-content/uploads/2025/14/pjs-25-1277-Figure-7.webp?">Figure 7</ext-link></p>
</fig>
<p>These results demonstrate that the model not only classifies accurately but also produces reliable probability estimates, which is essential for decision-making in safety-critical applications. The predicted probabilities (blue) closely follow the ideal diagonal (gray dashed), confirming good calibration.</p>
</sec>
<sec id="sec004-3">
<title>Ablation Study</title>
<p>To assess the contribution of each component in the proposed framework, an ablation study was performed. Results in <xref ref-type="table" rid="T5">Table 5</xref> show that while Random Forest and GNN individually achieved moderate accuracy, combining them significantly improved performance. The addition of SMOTE for class balancing and SHAP for interpretability further enhanced the hybrid model, demonstrating that each module contributed to the overall effectiveness of the system.SMOTE enhanced recall by addressing class imbalance, particularly for severe crashes. SHAP, while not improving accuracy, provided essential interpretability, making the framework more trustworthy for stakeholders.</p>
<table-wrap id="T5">
<label>Table 5</label>
<caption><title>Ablation Results</title></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left">Model Configuration</th>
<th valign="top" align="center">Accuracy (%)</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Random Forest only</td>
<td valign="top" align="center">90.64</td>
</tr>
<tr>
<td valign="top" align="left">GNN only</td>
<td valign="top" align="center">74.80</td>
</tr>
<tr>
<td valign="top" align="left">RF (PCA features only)</td>
<td valign="top" align="center">88.50</td>
</tr>
<tr>
<td valign="top" align="left">GNN + RF (no SMOTE)</td>
<td valign="top" align="center">91.50</td>
</tr>
<tr>
<td valign="top" align="left">GNN + RF (with SMOTE)</td>
<td valign="top" align="center">93.00</td>
</tr>
<tr>
<td valign="top" align="left"><bold>Full GNN + RF + SHAP</bold></td>
<td valign="top" align="center"><bold>94.20</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>For completeness, we also evaluated a Z-only configuration, where Random Forest was trained solely on PCA-encoded features without graph embeddings. This achieved 88.50% accuracy, confirming that PCA features alone provide moderate predictive power but fall short of the relational learning offered by GNN embeddings. The improvement from Z-only to GNN+RF demonstrates the added value of combining graph-based and tabular feature representations.</p>
<p>Additionally, the adoption of SMOTE for data resampling plays a pivotal role in ensuring balanced learning. By addressing the class imbalance, particularly for severe crash cases, SMOTE enables the model to generalize better across all severity levels. The ensemble generalization offered by combining GNN with RF allows the hybrid model to leverage both spatial and statistical feature learning strengths. This dual capacity results in a more accurate and generalizable system.<sup><xref ref-type="bibr" rid="ref13">13</xref></sup></p>
</sec>
<sec id="sec004-4">
<title>Expanded Baselines and Sensitivity Analysis</title>
<p>To avoid bias and capture all aspects of the problem, besides classical GCN and GAT architectures, baselines from the recent deep tabular learning field, e.g. Logistic Regression, TabNet, TabTransformer, and FT-Transformer, were considered. These baselines reflect both deep tabular learning and graph-specific models, thus enabling a more thorough comparison. Besides, sensitivity analyses were performed for several important hyperparameters: the number of neighbors in the kNN graph (k &#x2208; {3, 5, 10, 15}), the number of PCA components (30, 50, 70), GraphSAGE hidden dimensions (64, 128, 256), and SMOTE resampling ratios (1:1, 1:2, 1:3). In all experiments with various parameter settings, the hybrid GNN+RF was the best performer all the time achieving accuracy in the range of 92%&#x2013;94%. Thus it not only shows the advantage of the hybrid model over solo methods but also indicates the stability and the robustness, as well as the reliability for real-world applications, of the hybrid approach.</p>
</sec>
<sec id="sec004-5">
<title>Interpretability Analysis</title>
<p>To evaluate the transparency of the proposed framework, both permutation-based feature importance and SHAP analysis were conducted. SHAP was applied to both original features (RF) and GNN embeddings (RF on H). This ensured interpretability across raw features and learned representations</p>
<p><xref ref-type="fig" rid="F8">Figure 8</xref> shows global feature importance for the Random Forest on original features, where number of casualties, vehicles involved, and lighting conditions were dominant predictors.</p>
<fig id="F8" position="float">
<object-id pub-id-type="doi">10.70389/journal.PJS.100131.g008</object-id>
<label>Fig 8</label>
<caption><title>Global feature importance (RF on original features)</title></caption>
<p><ext-link ext-link-type="uri" xlink:href="https://i0.wp.com/premierscience.com/wp-content/uploads/2025/14/pjs-25-1277-Figure-8.webp?">Figure 8</ext-link></p>
</fig>
<p><xref ref-type="fig" rid="F9">Figure 9</xref> highlights important dimensions from the GNN embeddings, confirming that learned representations capture diverse aspects of crash data.</p>
<fig id="F9" position="float">
<object-id pub-id-type="doi">10.70389/journal.PJS.100131.g009</object-id>
<label>Fig 9</label>
<caption><title>Important embedding dimensions (RF on GNN embeddings)</title></caption>
<p><ext-link ext-link-type="uri" xlink:href="https://i0.wp.com/premierscience.com/wp-content/uploads/2025/14/pjs-25-1277-Figure-9.webp?">Figure 9</ext-link></p>
</fig>
<p><xref ref-type="fig" rid="F10">Figure 10</xref> presents SHAP global feature contributions, revealing consistent alignment with domain knowledge: casualty count, driver age, and road conditions significantly impacted severity predictions.</p>
<fig id="F10" position="float">
<object-id pub-id-type="doi">10.70389/journal.PJS.100131.g010</object-id>
<label>Fig 10</label>
<caption><title>Mean SHAP values (global impact across classes)</title></caption>
<p><ext-link ext-link-type="uri" xlink:href="https://i0.wp.com/premierscience.com/wp-content/uploads/2025/14/pjs-25-1277-Figure-10.webp?">Figure 10</ext-link></p>
</fig>
<p>Finally, SHAP interaction plots (<xref ref-type="fig" rid="F11">Figure 11</xref>) provide local interpretability, showing how specific factors such as driver age and time of accident jointly influence predictions.</p>
<fig id="F11" position="float">
<object-id pub-id-type="doi">10.70389/journal.PJS.100131.g011</object-id>
<label>Fig 11</label>
<caption><title>SHAP Interaction values (local interpretability)</title></caption>
<p><ext-link ext-link-type="uri" xlink:href="https://i0.wp.com/premierscience.com/wp-content/uploads/2025/14/pjs-25-1277-Figure-11.webp?">Figure 11</ext-link></p>
</fig>
<p>These results demonstrate that the hybrid GNN+RF not only achieves strong predictive accuracy but also produces interpretable outputs, allowing stakeholders to validate and trust model recommendations.</p>
<p>Another important observation is the hybrid model&#x2019;s superior convergence behavior and learning stability, as demonstrated by its smoother and more rapid improvement in accuracy across training epochs. Compared to the standalone GNN, the hybrid architecture shows reduced fluctuations in validation accuracy, suggesting better learning dynamics and less susceptibility to overfitting.</p>
</sec>
<sec id="sec004-6">
<title>Deployment and Fairness Analysis</title>
<p>For deployment in the real world, the hybrid GNN+RF model was also evaluated. The system showed effective training with an average runtime of about 14 minutes for the combined GraphSAGE and Random Forest framework, while inference was made in ~18 milliseconds per crash record (which is the same as 55 frames per second). The memory footprint for the process of training was measured at 2.3 GB of GPU utilization, thus confirming the possibility of a real-time deployment in traffic monitoring centers. Moreover, fairness analyses along subgroups gave lower recall rates for night-time crashes (&#x2013;3%) and motorcycle-related crashes (&#x2013;2.5%). These differences, which are quite small in size, nevertheless, pinpoint the possibility of bias in the model&#x0027;s prediction. To lessen these differences, the next step will be to work on the re-weighted loss functions and subgroup-specific resampling strategies so that accident performance can be more equitable in all kinds of scenarios.</p>
</sec>
</sec>
<sec id="sec005" sec-type="conclusions">
<title>Conclusion</title>
<p>The comparative evaluation conducted in this study highlights the strengths and limitations of standalone Graph Neural Networks (GNNs) in the context of road crash severity prediction. While GNNs, particularly those employing the GraphSAGE architecture, demonstrate an ability to capture relational patterns and contextual dependencies among crash instances, their standalone application falls short in achieving the robustness required for high-stakes safety-critical deployments.To overcome these limitations, we proposed a hybrid model that integrates GNN with Random Forest (RF), capitalizing on the spatial learning capabilities of GNN and the ensemble-based decision robustness of RF. This synergistic approach yielded a substantial improvement in performance, with the hybrid model achieving a validation accuracy exceeding 94%, as compared to the 74.8% attained by the GNN model alone. Additional improvements were observed across other evaluation metrics, including precision, recall, and F1-score, further affirming the effectiveness of the hybrid architecture.Moreover, the hybrid model exhibited faster convergence, better learning stability, and reduced fluctuations in accuracy across epochs. The confusion matrix analysis demonstrated its proficiency in classifying multiple severity levels with minimal misclassification, supporting its suitability for multi-class real-world applications.Given these findings, the hybrid GNN + RF model emerges as a compelling solution for intelligent traffic management systems. Its strength in merging relational reasoning accurate classification, and clear explanations makes it useful to apply in real-time crash prediction tools for emergency response, and planning infrastructure guided by policies. The model works well with Explainable AI (XAI), which boosts trust and clarity helping it gain wider approval from both professionals and stakeholders. In summary, this research shows how a hybrid learning approach blending graph-based representation with ensemble classification improves predictions while tackling issues like clarity and adaptability. Possible future work could look into using temporal and spatial GIS data trying it out in edge systems, or bringing it into fields like healthcare or urban risk analysis.</p>
<p>While the new framework exhibits excellent predictive power, its constructors still point out some of its limitations. The available data is, first of all, restricted to the UK, and it is unknown how well the model will perform outside this area. Secondly, the issue about deployment in real-time seems to be the lack of computational resources and the waiting time before the order is executed, thus the question whether it can run smoothly in these conditions still needs to be answered. Thirdly, even after applying the Synthetic Minority Oversampling Technique (SMOTE), the problem of residual class imbalance may linger in certain crash categories that are rare. Finally, the traditional technological and social problems such as fairness and accountability in decision-making are still a substantial source of concern and require comprehensive research to be solved.</p>
<p>These questions are going to be solved in future studies that will approach this problem using the richness of spatio-temporal data, trying out focal loss to get better handling of the imbalance, and deploying the optimized GNN models into the traffic monitoring scene in the field.</p>
<p>We put into practice the most demanding transparency measures, and in order for all this to be verifiable by everyone, we are now putting at your disposal the code, some configuration files, and the preprocessing scripts (feature encoding, missing data handling, SMOTE setup) used for the experiments. All the experiments were performed using fixed random seeds (42, 52, 62, 72, 82) on a computer with the following specs: Intel i7-11700K, 32 GB RAM, NVIDIA RTX 3080 (10 GB), Ubuntu 20.04, Python 3.10, PyTorch 2.2, PyTorch Geometric 2.5. The dataset can be accessed through the UK Department for Transport repository. The data preprocessing scripts guarantee that the results shown can be identically reproduced.</p>
</sec>
</body>
<back>
<fn-group>
<fn id="n1" fn-type="other">
<p>Additional material is published online only. To view please visit the journal online.</p>
<p><bold>Cite this as:</bold> Bhosale JB and Yelure BS. A Graph Based Hybrid Approach to Injury Severity Prediction in Road Accidents Using Deep Learning. Premier Journal of Science 2025;14:100131</p>
<p><bold>DOI:</bold> <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.70389/PJS.100131">https://doi.org/10.70389/PJS.100131</ext-link></p>
</fn>
<fn id="n2" fn-type="other">
<p><bold>Ethical approval</bold></p>
<p>N/a</p>
</fn>
<fn id="n3" fn-type="other">
<p><bold>Consent</bold></p>
<p>N/a</p>
</fn>
<fn id="n4" fn-type="other">
<p><bold>Funding</bold></p>
<p>No industry funding</p>
</fn>
<fn id="n5" fn-type="conflict">
<p><bold>Conflicts of interest</bold></p>
<p>N/a</p>
</fn>
<fn id="n6" fn-type="other">
<p><bold>Author contribution</bold></p>
<p>Jyoti B. Bhosale and Bhushan S. Yelure &#x2013; Conceptualization, Writing &#x2013; original draft, review and editing</p>
</fn>
<fn id="n7" fn-type="other">
<p><bold>Guarantor</bold></p>
<p>Jyoti B. Bhosale</p>
</fn>
<fn id="n8" fn-type="other">
<p><bold>Provenance and peer-review</bold></p>
<p>Unsolicited and externally peer-reviewed</p>
</fn>
<fn id="n9" fn-type="other">
<p><bold>Data availability statement</bold></p>
<p>N/a</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="ref1"><label>1</label><mixed-citation publication-type="journal"><string-name><surname>Ma</surname> <given-names>J</given-names></string-name>, <string-name><surname>Kockelman</surname> <given-names>K</given-names></string-name>. <article-title>Crash frequency and severity modeling using clustered data from Washington State</article-title>. <source>In: Proceedings of the IEEE Intelligent Transportation Systems Conference</source>; <year>2006</year>. p. <fpage>1621</fpage>&#x2013;<lpage>6</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/ITSC.2006.1707420">https://doi.org/10.1109/ITSC.2006.1707420</ext-link></mixed-citation></ref>
<ref id="ref2"><label>2</label><mixed-citation publication-type="journal"><string-name><surname>Tang</surname> <given-names>F</given-names></string-name>, <string-name><surname>Fu</surname> <given-names>X</given-names></string-name>, <string-name><surname>Cai</surname> <given-names>M</given-names></string-name>, <string-name><surname>Lu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Zhong</surname> <given-names>S</given-names></string-name>. <article-title>Investigation of the factors influencing the crash frequency in expressway tunnels: considering excess zero observations and unobserved heterogeneity</article-title>. <source>IEEE Access</source>. <year>2021</year>;<volume>9</volume>:<fpage>58549</fpage>&#x2013;<lpage>65</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/ACCESS.2021.3072539">https://doi.org/10.1109/ACCESS.2021.3072539</ext-link></mixed-citation></ref>
<ref id="ref3"><label>3</label><mixed-citation publication-type="journal"><string-name><surname>Ma</surname> <given-names>C</given-names></string-name>, <string-name><surname>Hao</surname> <given-names>W</given-names></string-name>, <string-name><surname>Xiang</surname> <given-names>W</given-names></string-name>, <string-name><surname>Yan</surname> <given-names>W</given-names></string-name>. <article-title>The impact of aggressive driving behavior on driver-injury severity at highway-rail grade crossing accidents</article-title>. <source>J Adv Transp</source>. <year>2018</year>;<volume>2018</volume>:<fpage>1</fpage>&#x2013;<lpage>10</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1155/2018/9841498">https://doi.org/10.1155/2018/9841498</ext-link></mixed-citation></ref>
<ref id="ref4"><label>4</label><mixed-citation publication-type="journal"><string-name><surname>Mesa-Arango</surname> <given-names>R</given-names></string-name>, <string-name><surname>Valencia-Alaix</surname> <given-names>VG</given-names></string-name>, <string-name><surname>Pineda-Mendez</surname> <given-names>RA</given-names></string-name>, <string-name><surname>Eissa</surname> <given-names>T</given-names></string-name>. <article-title>Influence of socioeconomic conditions on crash injury severity for an urban area in a developing country</article-title>. <source>Transp Res Rec</source>. <year>2018</year>;<volume>2672</volume>(<issue>31</issue>):<fpage>41</fpage>&#x2013;<lpage>53</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1177/0361198118780768">https://doi.org/10.1177/0361198118780768</ext-link></mixed-citation></ref>
<ref id="ref5"><label>5</label><mixed-citation publication-type="journal"><string-name><surname>Chen</surname> <given-names>C</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>G</given-names></string-name>, <string-name><surname>Huang</surname> <given-names>H</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>J</given-names></string-name>, <string-name><surname>Tarefder</surname> <given-names>RA</given-names></string-name>. <article-title>Examining driver injury severity outcomes in rural non-interstate roadway crashes using a hierarchical ordered logit model</article-title>. <source>Accid Anal Prev</source>. <year>2016</year>;<volume>96</volume>:<fpage>79</fpage>&#x2013;<lpage>87</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.aap.2016.07.033">https://doi.org/10.1016/j.aap.2016.07.033</ext-link>.</mixed-citation></ref>
<ref id="ref6"><label>6</label><mixed-citation publication-type="journal"><string-name><surname>Rifaat</surname> <given-names>SM</given-names></string-name>, <string-name><surname>Chin</surname> <given-names>HC</given-names></string-name>. <article-title>Accident severity analysis using ordered probit model</article-title>. <source>J Adv Transp</source>. <year>2007</year>;<volume>41</volume>(<issue>1</issue>):<fpage>91</fpage>&#x2013;<lpage>114</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1002/atr.5670410109">https://doi.org/10.1002/atr.5670410109</ext-link></mixed-citation></ref>
<ref id="ref7"><label>7</label><mixed-citation publication-type="journal"><string-name><surname>Liu</surname> <given-names>P</given-names></string-name>, <string-name><surname>Fan</surname> <given-names>W</given-names></string-name>. <article-title>Modeling head-on crash severity on NCDOT freeways: a mixed logit model approach</article-title>. <source>Can J Civ Eng</source>. <year>2019</year>;<volume>46</volume>(<issue>4</issue>):<fpage>322</fpage>&#x2013;<lpage>8</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1139/cjce-2018-0345">https://doi.org/10.1139/cjce-2018-0345</ext-link></mixed-citation></ref>
<ref id="ref8"><label>8</label><mixed-citation publication-type="journal"><string-name><surname>Azimi</surname> <given-names>G</given-names></string-name>, <string-name><surname>Rahimi</surname> <given-names>A</given-names></string-name>, <string-name><surname>Asgari</surname> <given-names>H</given-names></string-name>, <string-name><surname>Jin</surname> <given-names>X</given-names></string-name>. <article-title>Severity analysis for large truck rollover crashes using a random parameter ordered logit model</article-title>. <source>Accid Anal Prev</source>. <year>2020</year>;<volume>135</volume>:<fpage>105355</fpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1016/j.aap.2019.105355">https://doi.org/10.1016/j.aap.2019.105355</ext-link></mixed-citation></ref>
<ref id="ref9"><label>9</label><mixed-citation publication-type="journal"><string-name><surname>Shao</surname> <given-names>X</given-names></string-name>, <string-name><surname>Ma</surname> <given-names>X</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>F</given-names></string-name>, <string-name><surname>Song</surname> <given-names>M</given-names></string-name>, <string-name><surname>Pan</surname> <given-names>X</given-names></string-name>, <string-name><surname>You</surname> <given-names>K</given-names></string-name>. <article-title>A random parameters ordered probit analysis of injury severity in truck-involved rear-end collisions</article-title>. <source>Int J Environ Res Public Health</source>. <year>2020</year>;<volume>17</volume>(<issue>2</issue>):<fpage>395</fpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/ijerph17020395">https://doi.org/10.3390/ijerph17020395</ext-link></mixed-citation></ref>
<ref id="ref10"><label>10</label><mixed-citation publication-type="journal"><string-name><surname>Xie</surname> <given-names>S</given-names></string-name>, <string-name><surname>Ji</surname> <given-names>X</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>W</given-names></string-name>, <string-name><surname>Fang</surname> <given-names>R</given-names></string-name>, <string-name><surname>Hao</surname> <given-names>J</given-names></string-name>. <article-title>Exploring risk factors with crash severity on China two-lane rural roads using a random-parameter ordered probit model</article-title>. <source>J Adv Transp</source>. <year>2020</year>;<volume>2020</volume>:<fpage>1</fpage>&#x2013;<lpage>14</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1155/2020/8822852">https://doi.org/10.1155/2020/8822852</ext-link></mixed-citation></ref>
<ref id="ref11"><label>11</label><mixed-citation publication-type="journal"><string-name><surname>Kumeda</surname> <given-names>B</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>F</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>F</given-names></string-name>, <string-name><surname>Hussain</surname> <given-names>S</given-names></string-name>, <string-name><surname>Almasri</surname> <given-names>A</given-names></string-name>, <string-name><surname>Assefa</surname> <given-names>M</given-names></string-name>. <article-title>Classification of road traffic accident data using machine learning algorithms</article-title>. <source>In: Proceedings of the IEEE 11th International Conference on Communication Software and Networks (ICCSN)</source>; <year>2019</year>. p. <fpage>682</fpage>&#x2013;<lpage>7</lpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.1109/ICCSN.2019.8905302">https://doi.org/10.1109/ICCSN.2019.8905302</ext-link></mixed-citation></ref>
<ref id="ref12"><label>12</label><mixed-citation publication-type="journal"><string-name><surname>Lee</surname> <given-names>J</given-names></string-name>, <string-name><surname>Yoon</surname> <given-names>T</given-names></string-name>, <string-name><surname>Kwon</surname> <given-names>S</given-names></string-name>, <string-name><surname>Lee</surname> <given-names>J</given-names></string-name>. <article-title>Model evaluation for forecasting traffic accident severity in rainy seasons using machine learning algorithms: Seoul city study</article-title>. <source>Appl Sci</source>. <year>2020</year>;<volume>10</volume>(<issue>1</issue>):<fpage>129</fpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/app10010129">https://doi.org/10.3390/app10010129</ext-link></mixed-citation></ref>
<ref id="ref13"><label>13</label><mixed-citation publication-type="journal"><string-name><surname>Aldhari</surname> <given-names>I</given-names></string-name>, <string-name><surname>Almoshaogeh</surname> <given-names>M</given-names></string-name>, <string-name><surname>Jamal</surname> <given-names>A</given-names></string-name>, <string-name><surname>Alharbi</surname> <given-names>F</given-names></string-name>, <string-name><surname>Alinizzi</surname> <given-names>M</given-names></string-name>, <string-name><surname>Haider</surname> <given-names>H</given-names></string-name>. <article-title>Severity prediction of highway crashes in Saudi Arabia using machine learning techniques</article-title>. <source>Appl Sci</source>. <year>2023</year>;<volume>13</volume>(<issue>1</issue>):<fpage>233</fpage>. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3390/app13010233">https://doi.org/10.3390/app13010233</ext-link></mixed-citation></ref>
</ref-list>
</back>
</article>
