Skip to main content
PLOS One logoLink to PLOS One
. 2026 Apr 8;21(4):e0341961. doi: 10.1371/journal.pone.0341961

AI-assisted optimization design of seismic performance parameters for timber structures

Dongqi Wei 1, Yuqiang Ding 1, Feng Zhou 1, Xuan Zhang 2,*
Editor: Dajiang Geng3
PMCID: PMC13061333  PMID: 41950285

Abstract

Timber multi-story buildings offer environmental benefits, lightweight construction, and seismic resilience, but Artificial Intelligence (AI) based integrated frameworks for optimizing seismic parameters, including inter-story drift and roof displacement, remain limited. The Gradient Boosted Random Forest Machine with Scalable Cheetah Optimizer (GBRF-SCO) is proposed for improving prediction accuracy while facilitating optimal design decisions. The dataset consists of 4,000 timber building samples obtained from a publicly available Kaggle repository (Timber Seismic Performance Dataset). Data pre-processing employs normalization and outlier detection using Robust Scaling and Isolation Forest, ensuring high-quality inputs. For exploratory analysis, t-Distributed Stochastic Neighbor Embedding (t-SNE) is applied to visualize high-dimensional feature relationships and identify structural parameter patterns relevant to seismic performance. The proposed framework uses GBRF to predict seismic response metrics, with the SCO tweaking hyperparameters to optimize model performance. It also enables the optimization of seismic performance characteristics, guiding engineers in selecting structural designs that minimize drift and enhance robustness. Multiple Linear Regression (MLR) was employed to examine the influence of key structural and seismic elements on roof displacement, providing insights into the overall seismic performance of wood buildings. Comparative evaluation shows superior performance over conventional regression and ensemble methods, demonstrating a higher accuracy of 0.949, which corresponds to the classification of roof displacement levels (low, medium, high) under seismic loading conditions and seismic intensities using Python 3.10. By providing a strong and clever method for designing sustainable and earthquake-resilient buildings, the suggested GBRF-SCO framework successfully improves the seismic performance optimization of timber structures.

1. Introduction

Wood constructions play a vital role in modern building due to their exclusive grouping of strength, flexibility, and sustainability, promoting both builders and landlords of large-scale infrastructures [1,2]. Over the past 50 years, loans in engineering techniques have enabled wood to transition from a traditional framing material to a primary structural component in commercial and residential buildings. Engineered timber products enhance strength, stability, and fire resistance compared to conventional wood-framed structures, while offering a higher strength-to-weight ratio that facilitates faster construction and reduces environmental impact relative to concrete or steel [3,4]. Technological developments ensure timber remains a durable, adaptable, and sustainable material choice in contemporary design.

Earthquake performance is a critical consideration for timber buildings, as seismic forces can induce inter-story drift, roof displacement, base shear, acceleration, and energy dissipation [5,6]. Evaluating these performance indicators allows engineers to determine whether a structure possesses sufficient strength, stability, and energy-dissipation capacity to resist seismic events [7]. Key metrics include lateral displacement, inter-story drift, base shear, ductility, damping capacity, and the natural period of vibration, which collectively describe a timber building’s ability to withstand seismic loads without catastrophic failure. Adherence to earthquake-resistant design principles ensures human safety and structural resilience [8].

Engineers improve seismic performance by modifying structural systems, material selection, and construction details, while verifying that buildings are safe, serviceable, and compliant with codes [9,10]. Specific seismic performance parameters, essential for assessing timber structures, rely on intrinsic wood properties such as ductility, low mass, and flexibility [11,12]. Specifically, lateral displacement, inter-story drift, base shear, damping capacity, and energy dissipation are key indicators of timber structure resilience under seismic loading [13]. Moreover, the form, layout, and detailing of timber connections, as well as the use of engineered wood products, significantly influence seismic performance [14]. Proper assessment ensures timber structures can safely absorb and transfer seismic energy, limiting damage and protecting occupants. Conforming to current seismic codes and performance-based design standards requires careful consideration of these factors throughout the design and construction process [15]. Fig 1 shows the detailed structural framework of a timber building.

Fig 1. Structural framework of a timber building illustrating key components influencing seismic performance.

Fig 1

[Timber-Frames: Anatomy and Joinery – Custom Home Building and Remodeling].

It is crucial to monitor seismic performance, such as lateral drifts and roof displacement, to ensure that structures are safe under earthquake loads.

The research introduces the GBRF-SCO framework, an AI-driven methodology that enhances essential seismic metrics, including inter-story drift and roof displacement. Engineers can use the model’s hyperparameter optimization to make better predictions and design buildings that are more resistant to earthquakes and less likely to drift. In general, the GBRF-SCO framework helps with safer, stronger, and faster design of timber buildings.

1.1. Key contribution of the research

  • Initially, analytical modeling was carried out using a comprehensive Timber Seismic Performance Dataset that included different seismic and structural characteristics such as height, density, and acceleration.

  • After collecting the data, the dataset was processed through robust scaling, isolation forest, outlier removal, and normalization to enhance the model’s accuracy and reliability.

  • t-SNE analysis to visualize relationships between structural and seismic variables, identifying key trends for performance optimization.

  • Demonstration that the GBRF-SCO framework achieves superior predictive performance and stability, with improved accuracy, precision, and F1-score over conventional models.

2. Literature review

2.1. AI and machine learning in timber seismic performance

The research aimed to develop a multi-objective optimization framework [16] leveraging Deep Learning (DL) for the design of 20-story cross-laminated timber (CLT) Coupled Wall systems, improving high-dimensional design handling. An autoencoder compressed the design space dimensions, while neural networks mapped input variables to latent spaces and linked latent variables to output responses for structural optimization. This framework generated the best Pareto front, accounting for uncertainties in connection elements, and outperformed three deterministic models using nonlinear time history seismic analysis. However, it relied on a two-dimensional numerical model, limiting full three-dimensional representation, and its effectiveness depended on training data quality and diversity. Complementary research [17] applied machine learning (ML) algorithms to predict design suitability of box-shaped timber members, achieving accuracies from 91.7% to 98.6%, with Logistic Regression performing best. Both studies were limited by small datasets and constrained structural dimensions, potentially overlooking full structural variability.

2.2. Hybrid and CLT timber systems

The research [18] evaluated the mechanical performance of a hybrid steel-grout connection for CLT panels, using quasi-static cycle tests to assess secant stiffness and residual slip. ML models predicted these properties based on material and geometric characteristics, highlighting rod and grout diameters, though results are specific to the tested connection and large-scale CLT assemblies. Another study [19] employed DL with a deep fully convolutional neural network (d-FCNN) using encoder-decoder architecture to segment cracks in 501 images of Yingxian Pagoda timber components. The model, trained with a batch size of six over 100 epochs, achieved average accuracy but is limited by dataset size and generalization. Anomaly detection in civil structures [20] combined Transfer Learning (TL) with Extended Node Strength Network (ENSN) to identify Regions-of-Uninterest, validated via laboratory shaking table tests. Hybrid RC-timber systems [21] for high-rise seismic zones showed up to 38.7% seismic force and 30.6% base shear reduction, though limited to numerical analysis.

2.3. Multi-hazard and performance-based design approaches

The study [22] evaluated the effectiveness of hybrid walls and Impact-Resilient Double Concave Frictional Pendulum (IR-DCFP) bearings in enhancing seismic performance of Light-Frame Timber Buildings (LFTBs) under severe ground motions. Using incremental dynamic analyses, archetype Chilean LFTB models were assessed for collapse margin ratios and fragility curves. The combined hybrid walls and IR-DCFP solution minimized collapse probability at Maximum Considered Earthquake (MCE) levels and provided functional isolation despite low wall density and compact bearings. Multi-hazard design frameworks [23] were tested on steel-timber hybrid connections in 18- and 36-story case buildings using Direct Displacement-Based Design (DDBD), controlling seismic drift and wind-induced accelerations, though limited to analytical and numerical modeling. Another study [24] quantified cap beams’ effects on East Asian wood frames supported by stone bases, showing 42% reduction in energy dissipation, 81% increase in lateral load capacity, and 77% improvement in elastic stiffness, though applicability was limited to specific joint types and stress conditions. Table 1 provides a comparative review of notable timber-related seismic studies, emphasizing their aims, modeling methodologies, main discoveries, and current research constraints to identify knowledge gaps for future development.

Table 1. Overview of innovative research on seismic behavior of timber buildings.

Ref. Objective Model Key Results Limitations
[25] Assess seismic resistance of optimized timber structures Topology-optimized glulam braced frame High material efficiency; brittle response due to low redundancy Single configuration; no experiments
[26] Improve seismic stability of heritage timber FEM with buckling-restrained braces >50% tilt reduction; improved stiffness Single simulation; no validation
[27] Evaluate CLT seismic performance Base-isolated CLT with FPS High ductility; optimal seismic response Simplified numerical model
[28] Joint energy–seismic optimization Parametric simulation framework Strong energy–seismic interaction Limited sites; no real-world analysis
[29] Enhance wood feature detection DL-based line segment detection Improved accuracy and robustness Limited generalization
[30] Classify seismic damage ML-based damage classification High prediction accuracy Data- and region-dependent

2.4. Knowledge gap

This section exposes limits in current research on the seismic performance of wood structures, as well as a lack of integrated AI-based optimization frameworks. Table 2 provides a comparative review of important research gaps discovered in prior research on the seismic performance of wood structures, as well as a description of how the current work contributes to resolving them.

Table 2. Analysis of identified research gaps and research contributions.

Ref. Research Gap Contribution
[15] DL-based CLT studies lack 3D and experimental validation AI-optimized framework validated with real-world data for accurate 3D seismic prediction
[16] ML models limited by small datasets and fixed geometries Large, diverse dataset enables robust, generalizable seismic prediction
[18] Hybrid connector studies lack scalability Unified AI model integrates multiple connection systems for seismic optimization
[22] Retrofit assessments rely mainly on simulations Real seismic data validation ensures practical, reliable design

This systematic review highlights the evolution of timber seismic research, identifies limitations, and positions the GBRF-SCO framework as a robust AI-based solution addressing these critical gaps.

2.5. Distinction between existing AI-based approaches and the proposed GBRF-SCO framework

Previous AI-based seismic researches on timber structures mainly rely on single learning models with manually tuned or grid-searched hyperparameters, which often suffer from high computational cost and suboptimal convergence in high-dimensional design spaces. In contrast, the proposed GBRF-SCO framework integrates a hybrid Gradient Boosted Random Forest model with the Scalable Cheetah Optimizer for adaptive, population-based hyperparameter tuning. By coupling prediction and optimization within a unified learning loop, the framework enhances nonlinear modeling capability, robustness, and design-oriented seismic performance optimization.

3. Methodology

An optimization framework with AI help has been developed to reliably anticipate and improve seismic performance indicators of wood structures, such as inter-story drift and roof displacement, enabling engineers to design safer and more robust timber buildings under earthquake conditions. Fig 2 depicts the GBRF-SCO framework’s overall workflow for improving the seismic performance of timber structures, showing the processes of data collecting, preprocessing, feature analysis, model training, and prediction.

Fig 2. The completed flowchart for enhancing seismic performance parameters of timber structures.

Fig 2

3.1. Data acquisition

The Timber Seismic Performance Dataset contains detailed information on the structural and seismic characteristics of timber buildings. It considers building height, number of storys, wall thickness, material density, and damping qualities, as well as seismic inputs such as peak ground acceleration and spectrum acceleration. Table 3 represents the collection of sample data.

Table 3. Sample data from the timber seismic performance dataset.

Parameter Sample 1 Sample 2 Sample 3 Sample 4
Building Height (m) 36.21781 76.55 61.23958 51.90609
Number of Storys 11 6 7 5
Story Height (m) 3.91272 3.40562 2.6579 3.35902
Wall Thickness (m) 0.16088 0.5326 0.38316 0.45603
Material Density (kg/m³) 701.6804 496.9688 467.4328 557.7363
Modulus of Elasticity (Pa) 9.85E + 09 1.14E + 10 1.37E + 10 9.06E + 09
Damping Coefficient 0.02285 0.03804 0.01479 0.05187
Damping Ratio per Story 0.0332 0.02278 0.04835 0.03449
Floor Mass (kg) 74798.29 19779.63 77220.55 29065.37
Natural Frequency (Hz) 6.47606 7.36455 4.99899 3.67033
Peak Ground Acceleration (g) 0.24833 0.2076 0.35973 0.2887
Spectral Acceleration (m/s²) 2.70666 2.51013 3.1417 3.13824
Lateral Load Resisting Ratio 0.47871 0.28382 0.37719 0.78881
Inter-Story Drift (m) Timber Pegs Hybrid Connectors Hybrid Connectors Timber Pegs
Roof Displacement (m) Pile Pile Pile Pile
Building Height (m) Hip Hip Gable Hip
Number of Storys Shear Wall Moment Frame Shear Wall Moment Frame
Story Height (m) Symmetric Asymmetric Symmetric Asymmetric
Wall Thickness (m) 0.00246 0.00602 0.00492 0.00729
Material Density (kg/m³) 0.02655 0.03641 0.03649 0.03525

The Timber Seismic Performance Dataset referenced in this work is a secondary dataset publicly available through the Kaggle repository (https://www.kaggle.com/datasets/freshersstaff/timber-seismic-performance-dataset/data). This dataset has been used under the terms and conditions provided by Kaggle. The data are accessible without restriction to any researcher, and no identifiable human subjects or sensitive personal information are included in the dataset.

3.2. Data preparation approaches

Data preparation for the research included tasks such as normalizing the data, detecting outliers, and extracting features. By ensuring that all input features are uniformly scaled, normalization enhances the efficiency and stability of model training. Outlier detection eliminates odd or inconsistent data points, which enhances the integrity of the dataset on the whole and is even further verified by experts familiar with the domain. The t-SNE procedure is used to extract and visualize important structural patterns and relationships among variables, which gives a clearer presentation of seismic performance characteristics.

3.2.1. Normalization.

The Standard Scaler assumes that information is normally distributed within each component and scales them such that the distribution is centered around 0, with a standard deviation (SD) of 1. The element’s mean and SD are computed, then the component is scaled based on seismic performance parameters expressed in Equation (1).

Zscaled = xμσ (1)

Normalized data by subtracting the mean (μ) and dividing by the standard deviation (σ), ensuring all features are on a comparable scale for improved model performance.

3.2.2. Outlier detection.

The Isolation Forest algorithm is used to find anomalies in the seismic and structural dataset by looking for samples that are different from the rest. The contamination parameter is set to 0.05, which means that about 5% of the observations are likely to be outliers. Samples with anomaly scores below the decision threshold are considered anomalous and taken out of the dataset because they show physically inconsistent or severe seismic–structural pairings. This method cuts down on noise and bias, makes the data more reliable, and ensures that the samples kept are a true representation of how timber structures behave in genuine seismic conditions. This makes model training more stable and effective.

3.2.3. Feature analysis using t-distributed stochastic neighbor embedding (t-SNE).

This method enables feature extraction through dimensionality reduction, or the compression of data while preserving the connections between data points through the seismic performance criteria. Visualizations of correlations enable better decision-making by transforming convoluted data into its patterns and illuminating similarities, correlations, and clusters between points in an informative manner. By means of conditional probabilities, t-SNE augments standard geometrical distances to represent these similarities. The conditional probabilities  Oi|j, which are stated in Equation (2), express that similar data points wj and wi became individuals.

Oi|j=exp(||wjwi||22σj2)ljexp(||wjwl||22σj2) (2)

It calculates the conditional probability Oij, showing how similar point i is to point j. Here, wi and wj are their feature vectors, and σj controls how distance affects similarity. The numerator measures how close the two points are, while the denominator ensures all probabilities around j sum to one. This helps t-SNE identify clusters of similar data points. In the original space, the probabilities were specified by Equation (3).

Oj,i=(Oj|i+Oi|j)2m (3)

It defines the joint probability Oj, i, which represents the combined similarity between points i and j. It is calculated as Oj,i=Oji+Oij2m. Here, Oji  and Oij  are the conditional probabilities measuring how similar each point is to the other, and m is the overall data points. It ensures the similarity measure is symmetric, meaning the relationship between two points is treated equally in both directions. Where 2m represents the dataset’s size. The smooth indicator of an efficient number of neighbors represents the perplexity parameter that the t-SNE algorithm takes as an input, as shown in Equation (4).

Perp(Oj)=2G(Oj) (4)

It defines the perplexity of point Oj as Perp(Oj)=2G(Oj). G(Oj) represents the Shannon entropy of the probability distribution for point j. Perplexity indicates how broadly the probability distribution spreads across nearby points, effectively showing the close neighbors. A greater perplexity value means that more neighboring points are considered when measuring similarity. Here, G(Oj) is the bit value of the Shannon entropy  Oj, as shown in Equation (5).

G(Oj)=\nolimitsiOi|jlog2Oi|j (5)

It defines the Shannon entropy G(Oj) as G(Oj)=i Oijlog2Oij. In this equation, Oij  represents the conditional probability that point i is similar to point j. The entropy G(Oj) measures the uncertainty or spread of these probabilities. A higher entropy value indicates that the similarities are distributed more evenly among many points, while a lower value means that only a few points are strongly similar to j. Equation (6) defines the probabilities at low-dimensional rji using this distribution.

rji=(1+||zjzi||2)1lk(1+||zlzk||2)1 (6)

zj  and z are the low-dimensional representations of data points j and i. The numerator, (1+zjzi2)1, measures how close the two points have a higher similarity. The denominator ||zlzk|| ensures that all probabilities in the low-dimensional integral sum to one. The t-SNE technique determines the lower-dimensional projection of the input data zj as zl, therefore reducing the divergence among oji  and rji. After data preparation, the data is divided into two categories: testing (30%) and training (70%).

3.3. Gradient boosted random forest machine with scalable cheetah optimizer (GBRF-SCO) for accurate seismic predictions in timber structures

The GBRF-SCO framework employs an AI-based predictive ensemble combining gradient boosting and random forest to model nonlinear seismic–structural relationships. SCO adaptively optimizes GBRF hyperparameters using a cheetah-inspired search strategy, enhancing prediction accuracy, stability, convergence, and computational efficiency, thereby supporting the design of safer, more earthquake-resilient timber structures.

3.3.1. Gradient boosted random forest (GBRF) for accurate seismic response prediction.

The GBRF model merges aspects of gradient boosting and random forest methods. The model successfully captures the intricate, nonlinear connections between seismic and structural events. The linear combination of many weak learners increases overall prediction accuracy and generalization. This model is used for estimating inter-story drift and roof displacement in timber structures.

  • Gradient Boosting (GB) algorithm

The GB improves prediction accuracy by successively merging many weak learners to reduce residual errors, making it particularly efficient at capturing complicated nonlinear correlations between structural and seismic data. The GB architecture is represented in Fig 3.

Fig 3. The schematic representation of GB decision tree.

Fig 3

The GB algorithm can approximate the underlying function E(w)  given an input matrix w and a vector of molecular properties. This function maps the relationship between the molecular descriptor and the biological activity. The function E^(w) is constructed in an additive manner. Equation (7) is used in gradient boosting to combine multiple weak learners into a strong predictive model by gradually minimizing the overall error.

E^(w)=\nolimitsn=1Nσ*E^n(w) (7)

Here, is the total number of iterations, E^(w) represents the total error of the model, E^n(w)  is the error from the nth  iteration or tree, and σ  is the rate of learning that determines how much each new model contributes to the last prediction. After the first iteration, minimize the following goal given a loss function, which assesses the quality of predictions pi in relation to real readouts. Equation (8) defines the optimization step for the nth iteration in the gradient boosting process.

E^n=argminF(K(Z,On1On1On) (8)

E^n represents the minimized error function at iteration n, and K(Z, On1) denotes the loss function that calculates the variations among the predicted values On1 and the true outcomes Z. The term K(Z, On1)On1 is the gradient of the loss function with respect to the previous prediction, indicating the direction of steepest increase in error. By taking the negative gradient, the model updates in the direction that reduces the error most effectively. The argminF  expression means that the algorithm searches for the function F that minimizes this difference. In simple terms, this equation guides the model to adjust its predictions step by step to reduce the overall error during training.

  • Random Forest (RF) classifier

The RF enhances model resilience and generalization by combining findings from several Decision Trees (DTs), decreasing overfitting, and ensuring consistent predictions of seismic reactions in various timber structure configurations. The recommended binary classification strategy makes use of the RF algorithm. Following the training phase, RF builds a huge number of decision trees and produces a class with an average prediction. RF hyperparameters were changed using grid search. Fig 4 denotes the architecture of the RF classifier.

Fig 4. The general architecture of the RF classifier.

Fig 4

This research uses 1000 random trees with a maximum depth of 10, utilizing a 0.5 confidence vote method and the Gini impurity criterion. The Gini impurity can be determined as follows in equation (9):

H=\nolimitsi=1DO(i)*(1o(i)) (9)

Here, O(i) represents the data point probability that belongs to class i, and D is the overall classes. The term O(i)(1O(i)) measures how mixed or impure the data is for each class. When all data points belong to one class, H becomes zero, indicating perfect purity. Higher values of Hmean indicate greater uncertainty or diversity among the classes.

3.3.2. Traditional Cheetah Optimizer (TCO).

The TCO is a cheetah-inspired metaheuristic optimizing timber structure parameters to enhance seismic performance by balancing global exploration and local exploitation. It employs three strategies to avoid local optima. Fig 5 illustrates SCO phases: initial exploration, transition, attacking, and final convergence, systematically guiding the search toward optimal solutions. In the figure, Wj,iS0 represents the initial position of the cheetah in the search space, and Wj,iS1, Wj,iS2, Wj,iS3, and Wj,iS4 are its updated positions during different hunting phases. qj,i and q~j,i are random factors controlling movement, while αj,iS1 and βj,iS2 are adaptive coefficients that balance exploration and convergence. WA,iS2 denotes the best or attacking position guiding the cheetah toward the optimal solution.

Fig 5. The optimization process of SCO (a) the exploration phase, where cheetahs search for potential solutions, (b) transition phase moving toward promising regions, (c) the attacking phase with rapid convergence to the best solutions, and (d) depicts the final convergence phase, where the algorithm stabilizes around the optimal result.

Fig 5

Scalable Cheetah Optimizer (SCO) for Timber Structure Seismic Parameter Optimization: The Scalable Cheetah Optimizer (SCO) is a metaheuristic optimization algorithm designed for efficient hyperparameter tuning and seismic parameter optimization in timber structures. The SCO extends the traditional Cheetah Optimization (CO) framework by introducing adaptive control mechanisms, scalable step-size updates, and opposite learning concepts, ensuring high accuracy, rapid convergence, and improved stability in high-dimensional optimization problems.

Initialization of individuals: The optimization process begins with initializing the positions of cheetahs (candidate solutions) in a d-dimensional search space. The initial value of the  ith variable for the j-th individual, Wj,i, is defined as Equation (10)).

Wj,i=LBi+q·(UBiLBi),j=1,2,,r;i=1,2,,c (10)

where LBi and UBi are the lower and upper bounds of the i-th variable, r denotes a random number, c is the total number of variables, and q is a uniform random number between 0 and 1. This ensures a uniform random distribution of initial solutions across the defined search space.

Searching technique: During the search phase, cheetahs evaluate their environment, including prey location, cover, and their own status. Each individual’s position is updated according to the same initialization principle (Equation 11), maintaining randomness in early exploration. The step size αj,is is dynamically adapted at each iteration based on the current iteration ratio sS, the distance between candidate solutions, and random perturbations expressed in Equation (11).

αj,is={0.001sS(UBLB)l=10.001sS|Wt+b1b3+b2×[Was(j,i)Wjl]+b2b1+b2×[Ws(j,i)Wjl]Ws(j,i)|+qandl1  (11)

It describes the adaptive control parameter α(j, i)s, which adjusts the step size for each variable during the optimization process. When l=1, the value of α(j, i)s is determined by the ratio of the current iteration. The overall number of iterations S, multiplied by the variable range (UBLB). When l1, the value depends on the distance between the target position Wt, current position Ws(j, i), and other reference positions like Was(j,i) and W`jl. The constants b1, b2, and b3 control the weight of each term, while q adds a variation’s random factor.

Sitting-and-Waiting Model: The SCO also incorporates a sit-and-wait strategy, reflecting the cheetah’s behavior of remaining stationary to reduce unnecessary movements are expressed in Equation (12).

Wj,is+1=Wj,is (12)

Here, the variable remains unchanged between iterations, allowing the algorithm to conserve computational resources while maintaining candidate diversity.

Attacking Strategy: During the attacking phase, cheetahs adjust positions dynamically based on prey, nearby individuals, and the leader’s position. The position update is expressed as Equation (13 and 14).

Wj,is+1=WA,is+q¯j,i·βj,is (13)
βj,is=W1s(j,i)Ws(j,i) (14)

where WA,is is the leader or reference position, q¯j,i is a random scaling factor, and βj,is represents the difference between the best solution and the current candidate. This allows efficient exploitation of promising areas in the search space.

SCO for Hyperparameter Optimization: The SCO framework ensures fast and reliable convergence in predicting seismic parameters by dynamically adjusting GBRF hyperparameters. Unlike the traditional TCO algorithm, which suffers from slow convergence, early stagnation, and inefficiency in high-dimensional problems, SCO is adaptive, scalable, and provides improved optimization performance. The opposite learning strategy further enhances exploration in Equation (15).

W~jl=γ·(k+v)Wjl (15)

where W~jl is the opposite position of the j-th individual in dimension l, γ is a scaling factor, and k, v defines the exploration center.

Control Parameter Update Strategy: The CO algorithm updates control parameters adaptively to match nonlinear optimization requirements. The step size αj,is is updated as Equation (16 and 17).

αj,is=0.001·sS(UBLB)for l=1 (16)
βj,is=Wt+b3b3+b2·(W1s(j,i)W¯jl)+b2b3+b2·(Ws(j,i)W¯jl),l1 (17)

Here, Wt is the target position, W1s(j,i) the best-known solution, and W¯jl the reference position. Constants b2 and b3 control the contribution of each term, while q introduces randomness. This strategy balances global exploration with local fine-tuning.

Updating Individual Locations: To enhance individual updates and prevent redundant searches, the SCO employs the following update mechanism for binary and probabilistic optimization is expressed in Equation (18).

Wj,is+1=11+exp(W¯j,is+1) (18)

The sigmoid function bounds updated values between 0 and 1, suitable for probabilistic decisions. For discrete optimization is compute as Equation (19).

Wj,is+1={1,q>W¯j,is+11,otherwise  (19)

This converts continuous positions to binary states, enabling SCO to handle both continuous and discrete seismic optimization problems effectively. In simple terms, this decides whether each variable can take the value –1 or 1 in the next iteration. The SCO algorithm employs initialization, searching, and attacking strategies with adaptive control, enhancing convergence, stability, and scalability while optimizing GBRF hyperparameters, improving accuracy and efficiency in seismic parameter optimization for timber structures. Table 4 defines the key hyperparameters and their mathematical formulations used in the GBRF-SCO for seismic performance prediction.

Table 4. Hyperparameter for GBRF and SCO optimization framework.
Category Hyperparameters Variables Values
GBRF Number of Trees N 50–300
Learning Rate σ 0.01–0.3
Maximum Depth d 2–8
Loss Function K(Z,O) Mean Squared Error (MSE)
Error Function E^(w) Computed value
SCO Population Size r 8–10
Maximum Iterations S 15–20
Step Size α (0.001 ×(s/ S))
Position Difference β Dynamic
Opposite Position Factor γ 0.5
Lower Bound LB Defined per variable
Upper Bound UB Defined per variable

The selection of SCO hyperparameters depends on how well they converge, how fast they can be computed, and how many dimensions the seismic optimization problem has. The population size (r = 8–10) makes sure that there are enough different solutions without making the computation too expensive in feature spaces with a lot of dimensions. Smaller populations limit research, but bigger ones provide slight advances in accuracy. Convergence analysis, which is when the optimization error stabilizes, tells you the maximum number of iterations (S = 15–20). This setup strikes a good mix between exploration and exploitation, which makes GBRF hyperparameter tuning steady and dependable.

Pseudo code 1 represents the GBRF-SCO framework, which integrates gradient boosting for learning seismic behaviour with the SCO to adaptively tune model parameters, achieving accurate and efficient seismic performance prediction for timber structures.

Pseudo Code 1: GBRF–SCO Framework for Seismic Performance Prediction in Timber Structures

Input:

X=[x1,x2,,xm] # Structural and seismic features

Y=[y1,y2,,ym] # Continuous seismic response (roof displacement / drift)

P # SCO population size

S # SCO iterations

LB, UB # Hyperparameter bounds

Output:

Y^final # Optimized seismic response prediction

Step 1: Data Preprocessing

1. Remove outliers using Isolation Forest

2. Normalize features using Robust Scaling

3. Split data into training and testing sets

Step 2: Gradient Boosted Random Forest (GBRF) Modeling

1. Initialize GBRF hyperparameters:

θ={nestimators,η,dmax)

2. Train GBRF model on (Xtrain,Ytrain)

3. Predict seismic response Y^

Step 3: Scalable Cheetah Optimizer (SCO)

1. Initialize the cheetah population:

Wj0=LB+rand(0,1)×(UBLB)

2. Fors=1to S:

Evaluate fitness:

Fitness(Wjs)=RMSE(Y,Y^)

Identify the best cheetah W*

Update position:

Wjs+1=σ(γ(k+v)Wjs+αs(W*Wjs))

Apply boundary constraints

3. Return optimal hyperparameters θ*

Step 4: Final Prediction

1. Retrain GBRF using θ*

2. Generate optimized seismic response: Y^final

Step 5: Seismic Damage Classification (Optional)

Class={Low,Y^final<Q33Medium,Q33Y^final<Q66High,Y^finalQ66

Return: Y^final

The GBRF-SCO framework partitions data, learns structural–seismic relationships by minimizing errors, and optimizes hyperparameters via cheetah-inspired SCO. Repeated learning identifies optimal parameters, while MLR assesses key factors, enabling the retrained GBRF to predict accurate, reliable seismic performance in timber structures.

4. Performance analysis and discussion

The objective is to provide an AI-assisted optimization framework for precisely predicting and improving the seismic performance characteristics of wood buildings, with an emphasis on inter-story drift and roof displacement. Table 5 presents the hardware and software specifications used for implementing and testing the GBRF-SCO framework.

Table 5. Experimental setup for GBRF-SCO model implementation.

Category Specification Details
Hardware Configuration Processor Intel Core i7, 3.4 GHz
RAM 16 GB DDR4
ROM/ Storage 512 GB SSD
System Type 64-bit Operating System
Software Configuration Operating System Windows 11/ Ubuntu 22.04
Programming Language Python 3.10
Development Environment Jupyter Notebook/ VS Code
Libraries Used NumPy, Scikit-learn, Matplotlib, Pandas
Optimization Framework SCO implemented in Python

4.1. Multiple linear regression analysis for evaluating seismic performance relationships

The MLR approach analyzes the effects associated with different structural and seismic indicators, such as building height, wall thickness, and the acceleration of motion on the roof displacement of timber buildings. Table 6 presents the MLR results showing how key structural and seismic parameters influence roof displacement and GBRF-SCO’s overall seismic performance in timber structures.

Table 6. MLR analysis of seismic parameters.

Independent Variable T Statistic Standard Error p-Value Significance
Intercept 3.95 0.0021 0.0001 ***
Building Height 5.25 0.00004 0.0000 ***
Number of Storys 4.22 0.00027 0.0000 ***
Wall Thickness −2.94 0.00182 0.0035 **
Material Density −2.00 0.000002 0.0470 *
PGA 4.30 0.0102 0.0000 ***
SA 2.81 0.0031 0.0052 **
Lateral Load Resisting Ratio −2.16 0.0058 0.0320 *
Damping Coefficient −2.77 0.0315 0.0063 **
Model Summary
R² = 0.921 p < 0.001 F (8, 120) = 56.82 Durbin–Watson = 1.95

Note: The MLR model achieved excellent fit (R² = 0.921), explaining 92.1% of roof displacement variation. Roof displacement increased with Building Height and PGA, while greater Wall Thickness and Damping Coefficient reduced it, with significance levels indicated by (p < 0.001), ** - moderately significant predictors (p < 0.01), * - significant predictors (p < 0.05),

4.2. Distribution of roof displacement by structural system type

Determine whether there are structural system-specific differences in the movement of roofs subjected to seismic forces. In order to determine which system can withstand an earthquake better in terms of stability and resilience. Better knowledge of structural behavior and more efficient design of wood structural systems lead to stronger buildings. Fig 6 shows the distribution of roof displacement among structural system types under seismic loading.

Fig 6. Roof displacement variation for timber structural system types.

Fig 6

These findings suggest that the GBRF-SCO model may provide a good representation of the structural system-specific variation in roof displacement. By consistently outperforming the others in displacement, the hybrid system proves that the suggested framework improves the precision of seismic performance forecasts.

4.3. Comparative analysis of seismic ground and structural parameters

The research examines the consequences that various seismic intensity parameters have on the lateral load resistant capabilities of varying buildings, focusing on the influence these parameters have on the stability, strength, and seismic performance of residential structures, which would assist in optimizing structural performance for diverse ground motions, allowing more factual assessment of overall earthquake resistance, depicted in Fig 7 for each structural sample under seismic acceleration characteristics and lateral load resisting ratios.

Fig 7. Stacked Area Plot of Seismic Parameters.

Fig 7

The patterns that overlap indicate a high correlation between the intensity of ground motion and the load-resisting capability of the structure. There is a little positive correlation between building height and roof displacement, as shown in Fig 8(a) by the trend line. Fig 8(b) Roof displacement versus peak ground acceleration is shown, with materials color-coded according to density. Fig 8(c) shows the overall patterns of fluctuation in roof displacement as a function of data indices, including the rolling mean variation.

Fig 8. Findings of (a) Relationship between building height and roof displacement, (b) Variation of roof displacement with peak ground acceleration, (c) Trend of roof displacement across samples showing overall fluctuations patterns.

Fig 8

The findings suggest that there is a connection between essential seismic parameters and roof displacement. The maximum ground acceleration varies from lightly felt to moderately felt to strongly felt; this is also true for roof displacement, which experiences a slight increase with greater height. The overall oscillation trends of the series of timber building samples reinforce some general consistent structural performance and stability of the GBRF-SCO.

4.4. Comparative analysis of roof displacement and structural feature relationships for different roof types

To investigate how roof geometry and critical structural characteristics affect roof displacement and overall seismic behavior, the study analyzes their influence on structural stability, deformation patterns, and the seismic performance of timber buildings. Fig 9(a) shows the variation of roof displacement with the number of storys for different roof types, indicating that displacement generally increases with height. Fig 9(b) illustrating relationships among structural features and roof displacement.

Fig 9. Graphical findings of (a) roof displacement with number of storys for different roof types, and (b) Parallel coordinate analysis of structural features influencing roof displacement across rooftypes.

Fig 9

The results show that the geometry of the roof has a major impact on the earthquake reaction, especially for sawtooth and hip roof types, and that more storys result in more displacement of the roof. Roof displacement is also significantly affected by wall thickness and material density, according to the study, which shows how geometry and material parameters work together to affect seismic performance.

4.5. Pairwise correlation analysis of key structural and seismic parameters

The examination is to explore interrelationships among structural parameters such as height, storys, wall thickness, and material density, and their combined influence on roof displacement under seismic loading. Fig 10 showing relationships between building and material parameters with roof displacement, helping identify key factors influencing seismic performance.

Fig 10. Pairwise correlation analysis of key structural and seismic parameters.

Fig 10

The findings indicate that roof displacement rises modestly with building height and number of storys, although wall thickness and material density have less association in GBRF-SCO. This suggests that geometric characteristics have a stronger effect on seismic reaction than material attributes in wood constructions.

4.6. Assessment of interdependencies among seismic response variables

Roof displacement and inter-story drift are two seismic and structural variables that the GBRF-SCO hopes to shed light on. These variables include building height, material properties, and ground acceleration. The correlation heatmap shown in Fig 11 illustrates the strength and direction of connections among structural, material, and seismic elements.

Fig 11. Correlation analysis of seismic and material parameters in timber structures.

Fig 11

There is a considerable positive connection between building height (0.35–0.54) and the number of storys (0.66), and the results show that there is a strong association (0.67) between inter-story drift and roof displacement.

4.7. Assessment of inter-story drift and roof displacement consistency trends

To assess the patterns of structural deformation, compare the inter-story drift with the movement of the roof. This shows that the structure is generally stable and resilient, which is consistent with the GBRF-SCO’s evaluation of the seismic forces’ propagation through wood structures. See how inter-story drift compares to roof deformation in wood buildings in Fig 12.

Fig 12. Comparison of roof displacement and inter-story drift in timber structures.

Fig 12

The result illustrates that the values of roof displacement are consistently greater than those of inter-story drift, suggesting that lateral deformation accumulates along the building’s height under seismic loading conditions.

4.8. Classification performance of roof displacement categories using a confusion matrix

The investigation verifies that the model accurately predicts the degree of roof displacement, which is necessary for correctly categorizing structural seismic reactions. Fig 13 displays the results of the roof displacement classes’ classification performance.

Fig 13. Confusion matrix representation of roof displacementclassificarion results.

Fig 13

Based on quantile thresholds from the dataset’s empirical distribution, roof displacement values are divided into three ordinal classifications such as low, medium, and high. Values of displacement that are less than the 33rd percentile are considered low, values that are between the 33rd and 66th percentiles are considered medium, and values that are greater than the 66th percentile are considered high. This classification, based on quantiles, keeps the class proportions even, keeps the ordinal structure of seismic reaction severity, and doesn’t let people choose the threshold. The established criteria provide a uniform interpretation of the confusion matrix and elucidate the lack of significant misclassifications between low and high displacement categories.

A small number of incorrect classifications show that the model can distinguish between wood structures with varying degrees of seismic reactivity when using GBRF-SCO. During seismic parameter prediction, Fig 14 shows that the model learned steadily and converged strongly.

Fig 14. Performance analysis of GBRF-SCO model across training epochs.

Fig 14

According to the results, the suggested GBRF-SCO model steadily increased testing and training accuracy across all epochs. To make the accuracy and loss trends easier to see, the legend was moved outside the graph. The GBRF-SCO model achieved a final training accuracy of 94.9% by displaying strong convergence and little validation variance.

4.9. Comparative analysis

This experiment evaluates the accuracy and efficiency of prediction models for seismic performance in timber structures, comparing the proposed Gradient Boosting Random Forest with Scalable Cheetah Optimizer (GBRF-SCO) to existing methods, including Enhanced Deep Line Segment Detection (LSD) [29], Support Vector Machine (SVM) [30], Adaptive Boosting (AdaBoost) [30], Extreme Gradient Boost (XGBoost) [30], Stacking Artificial Neural Network (ANN) [30], and Gradient Boosting ANN [30]. Performance is assessed using R², Mean Absolute Error (MAE), Root Mean Squared Error (RMSE), Accuracy, ROC-AUC, Precision, Recall, and F1-Score, ensuring a comprehensive evaluation of predictive fit, error magnitude, classification correctness, and positive prediction balance. Table 7 shows the numerical outcomes of error metrics.

Table 7. Evaluation of predictive modelling approaches with error metrics.

Approaches R2 MAE RMSE
Enhanced deep LSD [29] 0.98 0.68 0.23
GBRF-SCO [Proposed model] 0.988 0.54 0.20

The proposed GBRF-SCO model performed best with R² (0.988), MAE (0.54), and RMSE (0.20). The Enhanced Deep LSD model showed lower performance with R² (0.98), MAE (0.68), and RMSE (0.23). Overall, the proposed model gives more accurate and reliable results in Table 8. Fig 15 shows the Benchmark assessment of GBRF-SCO classification algorithms using quantitative performance metrics.

Table 8. Comparison of different classification model performances.

Approaches Accuracy Precision Recall F1-Score ROC AUC Score
SVM [30] 0.866 0.863 0.840 0.850 0.975
AdaBoost [30] 0.878 0.875 0.856 0.864 0.980
XGBoost [30] 0.882 0.874 0.863 0.868 0.982
Stacking ANN [30] 0.870 0.863 0.848 0.854 0.977
Gradient boosting ANN [30] 0.873 0.866 0.852 0.858 0.979
GBRF-SCO [Proposed model] 0.949 0.889 0.884 0.897 0.988

Fig 15. Evaluation of classification model accuracy across multiple performance indicators.

Fig 15

To reduce bias and have a fair comparison, all the ML models used hyperparameter settings taken directly from their original and peer reviewed studies. The only change made on the testing was the data set used for training. The data before and after counting, testing data, and tools to analyze the counting were the same. Such all-in-one testing method guarantees each model’s performance relies only on its skill and not on anything else apart from the model. The hyperparameter configurations for all the benchmark models used in the comparative analysis were used as they were in their original and peer-existing research to make the fair and unbiased comparison. Enhanced Deep LSD model makes use of best line confidence thresholds, crack length filter, and CNN based features extractor. The SVM makes use of polynomial kernel and a regularization parameter C = 45. AdaBoost makes use of a decision tree base estimator (max depth = 10), 180 estimators and learning rate of 1.0. The calculator uses a gbtree booster with β = 0.1, 450 estimators, Lambdas = 1.0 and Ï« = 0.01 with subsampling ratios of 0.7. The Stacking ant neural network integrated multiple ANN base learners to a meta learner of logistic regression (C = 20). The Gradient boosting NN made use of 25 boosting iterations with a learning rate of 0.2 and the weak learners was an ANN with two hidden layers with 128 neurons each. The proposed GBRF-SCO hyperparameters are fine-tuned by the SCO, which allows for an adaptive adjustment of the hyperparameters for better seismic performance estimation. All models in general were analysed under the same data preparation and experimental conditions so that every model could be objectively compared and their use could be re-produced.

The GBRF-SCO model outperformed all other methods, achieving the highest overall accuracy of 0.949, according to the performance comparison of several classification models. Along with the best ROC AUC Score (0.988), it had excellent performance across all other measures as well, including Precision (0.889), Recall (0.884), and F1-Score (0.897). The suggested model outperforms conventional models in terms of GBRF-SCO’s predictive power and class separation efficiency. Table 9 shows the results of the ablation studies that compared the efficiency of the individual and combined models.

Table 9. Ablation research results for model performance comparison.

Model Precision F1-Score Accuracy Recall
GBRF 0.861 0.857 0.902 0.854
SCO 0.872 0.870 0.918 0.868
GBRF-SCO (Proposed Model) 0.889 0.897 0.949 0.884

This ablation research showed that the combined GBRF-SCO model outperformed its individual components, GBRF and SCO. This integration led to higher accuracy, improved prediction stability, and better overall seismic performance optimization.

4.10. Discussion

The safety and resilience of wood buildings during earthquakes were enhanced through the development of the GBRF-SCO, capable of accurately forecasting and optimizing seismic performance parameters. Previous deep learning (DL) models for wood material identification showed promise but were dataset-specific and lacked generalizability [29]. Similarly, machine learning (ML) effectively assesses seismic damage in reinforced concrete (RC) structures, though its accuracy depends heavily on data quality and regional calibration [30]. Addressing these limitations, the study introduces the GBRF-SCO model, integrating AI-based prediction and optimization to improve seismic performance accuracy and design efficiency across diverse wood structure types.

4.10.1. Practical implications.

Engineers can make safer, more resilient buildings by designing wood structures with excellent seismic properties using the GBRF-SCO framework. It offers a data-driven decision-support tool that real construction projects can use to make buildings more seismically compliant and safer.

5. Conclusion

An AI-assisted GBRF-SCO framework was proposed for accurately predicting and optimizing key seismic performance parameters of timber structures. By integrating GBRF-SCO, the framework effectively captures complex nonlinear structural-seismic relationships while adaptively optimizing model hyperparameters. The proposed approach demonstrates clear performance advantages over conventional machine learning and ensemble models, achieving an R² of 0.988, MAE of 0.54, and RMSE of 0.20. Beyond predictive accuracy, the GBRF-SCO framework provides practical engineering value by supporting data-driven design decisions aimed at reducing inter-story drift and roof displacement, thereby enhancing the safety, resilience, and sustainability of timber buildings in seismic regions.

5.1. Limitations and future work

The GBRF-SCO framework has robust predictive and optimization capabilities; nevertheless, the study is limited by data diversity and availability, since the dataset predominantly depends on recorded and simulation-based data rather than comprehensive real-world experimental observations. The performance of the model depends on how good and complete the structural and seismic input parameters are. Also, the SCO-based optimization makes it more expensive to compute for high-dimensional feature spaces. Furthermore, extrapolation during intense seismic occurrences or atypical timber structural systems may be constrained by inadequate representative samples. Future efforts will focus on comprehensive experimental testing, field-based sensor data collection, multi-hazard scenario integration, and enhancements in computing efficiency to bolster resilience and practical application across various timber structures.

Supporting information

S1 File. We have uploaded the minimal dataset underlying the findings of this study as Supporting Information with the submission.

In addition, the dataset is publicly available without restriction from the Kaggle repository at the following link: https://www.kaggle.com/datasets/freshersstaff/timber-seismic-performance-dataset/data

(XLSX)

pone.0341961.s001.xlsx (665KB, xlsx)

Data Availability

All relevant data are within the manuscript.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Ilgın HE. Analysis of the main architectural and structural design considerations in tall timber buildings. Buildings. 2023;14(1):43. doi: 10.3390/buildings14010043 [DOI] [Google Scholar]
  • 2.Karaca F, Tleuken A, Pineda-Martos R, Ros Cardoso S, Orel D, Askar R, et al. Cultivating sustainable construction: stakeholder insights driving circular economy innovation for inclusive resource equity. Buildings. 2024;14(4):935. doi: 10.3390/buildings14040935 [DOI] [Google Scholar]
  • 3.Al-Najjar A, Dodoo A. Modular multi-storey construction with cross-laminated timber: life cycle environmental implications. Wood Mater Sci Eng. 2022;18(2):525–39. doi: 10.1080/17480272.2022.2053204 [DOI] [Google Scholar]
  • 4.Bergsagel D, Heisel F, Owen J, Rodencal M. Engineered wood products for circular construction: a multi-factor evaluation of lamination methods. npj Mater Sustain. 2025;3(1). doi: 10.1038/s44296-025-00067-7 [DOI] [Google Scholar]
  • 5.Yan L, Klingner R, Al-Qudsi A, Chen H, Dand JA. Current market landscape and industry voices in key timber construction markets. Buildings. 2025;15(18):3381. doi: 10.3390/buildings15183381 [DOI] [Google Scholar]
  • 6.De Araujo V. Timber construction as a multiple valuable sustainable alternative: main characteristics, challenge remarks and affirmative actions. Int J Constr Manag. 2021;23(8):1334–43. doi: 10.1080/15623599.2021.1969742 [DOI] [Google Scholar]
  • 7.Ilgın HE, Aslantamer ÖN. High-rise timber offices: main architectural and structural design parameters. Buildings. 2024;14(7):1951. doi: 10.3390/buildings14071951 [DOI] [Google Scholar]
  • 8.Vladimirova E, Gong M. Advancements and applications of wood-based sandwich panels in modern construction. Buildings. 2024;14(8):2359. doi: 10.3390/buildings14082359 [DOI] [Google Scholar]
  • 9.Shareef SS. Earthquake consideration in architectural design: guidelines for architects. Sustainability. 2023;15(18):13760. doi: 10.3390/su151813760 [DOI] [Google Scholar]
  • 10.Yang Z, Liu D, Zhou J, Zhang L, Liao W. Seismic vulnerability assessment of the cliff-attached buildings equipped with energy dissipation devices under obliquely incident seismic waves. Buildings. 2024;14(11):3488. doi: 10.3390/buildings14113488 [DOI] [Google Scholar]
  • 11.Zaker Esteghamati M. Leveraging machine learning techniques to support a holistic performance-based seismic design of civil structures. Woodhead Publishing; 2024. pp. 25–49. doi: 10.1016/b978-0-12-824073-1.00008-3 [DOI] [Google Scholar]
  • 12.Yang N, Qian X. Prediction of monitoring data and state assessment of heritage stone-timber structures based on spatio-temporal graph neural networks. Eng Struct. 2025;338:120619. doi: 10.1016/j.engstruct.2025.120619 [DOI] [Google Scholar]
  • 13.Liu Y, Xie J, Zhang W, Yu H. Research on hysteretic performance prediction method of novel connection dampers for external wall panels based on the mining of temporal combination characteristics. Structures. 2025;78:109285. doi: 10.1016/j.istruc.2025.109285 [DOI] [Google Scholar]
  • 14.Yuanyao M, Dongbo L, Chunyan L, Yan W, Xiguang L, Bo W. Multi-objective optimization of traditional residential timber frames based on response surface methodology. J Comput Methods Sci Eng. 2024;24(6):3493–503. doi: 10.1177/14727978241293251 [DOI] [Google Scholar]
  • 15.Lyu H, Oshio H, Matsuoka M. Deep learning-based collapsed building mapping from post-earthquake aerial imagery. Remote Sens. 2025;17(17):3116. doi: 10.3390/rs17173116 [DOI] [Google Scholar]
  • 16.Das S, Teweldebrhan BT, Tesfamariam S. High-dimensional multi-objective optimization of coupled cross-laminated timber walls building using deep learning. Eng Appl Artif Intell. 2024;136:109055. doi: 10.1016/j.engappai.2024.109055 [DOI] [Google Scholar]
  • 17.Cosut M, Bekdas G, Nigdeli SM, Isikdag U. Predicting design suitability of box-shaped sustainable timber structural members using machine learning and hyperparameter optimization. Neural Comput Applic. 2025;37(29):24123–48. doi: 10.1007/s00521-025-11556-0 [DOI] [Google Scholar]
  • 18.Feujofack K BV, Loss C. Experimental and machine learning study on a novel high-performance hybrid steel-grout connector for cross-laminated timber panels under pre-yield cyclic loads. Eng Struct. 2024;303:117530. doi: 10.1016/j.engstruct.2024.117530 [DOI] [Google Scholar]
  • 19.Zhang L, Xie Q, Wang H, Han J, Wu Y. Deep‐learning‐based crack identification and quantification for wooden components in ancient chinese timber structures. Struct Control Health Monit. 2024;2024(1). doi: 10.1155/2024/9999255 [DOI] [Google Scholar]
  • 20.Wang S, Saida T, Nishio M. Optical flow‐based structural anomaly detection in seismic events from video data combined with computational cost reduction through deep learning. Struct Control Health Monit. 2025;2025(1). doi: 10.1155/stc/4702519 [DOI] [Google Scholar]
  • 21.Tsai M-T, Hsu C-C. Assessment of structural performance, materials efficiency, and environmental impact of multi-story hybrid timber structures in high seismic zone. Case Stud Const Mater. 2024;21:e03695. doi: 10.1016/j.cscm.2024.e03695 [DOI] [Google Scholar]
  • 22.Quizanga D, Almazán JL, Torres-Rodas P, Guindos P. Seismic performance of timber buildings retrofitted with hybrid walls and impact-resilient isolators. Soil Dyn Earthq Eng. 2025;198:109586. doi: 10.1016/j.soildyn.2025.109586 [DOI] [Google Scholar]
  • 23.Ciabattoni M, Petrini F, Pampanin S. Multi-hazard design of low-damage high-rise steel–timber buildings subjected to wind and earthquake loading. Eng Struct. 2024;303:117522. doi: 10.1016/j.engstruct.2024.117522 [DOI] [Google Scholar]
  • 24.Wu Y-J, Liu C, Wang L, Xie Q-F, Zhang L-P. Seismic performance of stone supported traditional timber frames with or without cap beams. J Build Eng. 2023;70:106395. doi: 10.1016/j.jobe.2023.106395 [DOI] [Google Scholar]
  • 25.Wu Y-J, Zhang X-Y, Xie Q-F, Zhang L-P. Evaluation of seismic performance of traditional heavy timber frames with different types of column foot joints. J Build Eng. 2025;103:112123. doi: 10.1016/j.jobe.2025.112123 [DOI] [Google Scholar]
  • 26.Solis F, Parra PF, Cendoya P, Gonzalez-Böhme LF, Quitral-Zapata F, Gallardo R. Seismic Behavior of a Timber Structure Based on a Soft-Kill BESO Optimization Algorithm. Buildings. 2025;15(6):980. doi: 10.3390/buildings15060980 [DOI] [Google Scholar]
  • 27.Auad G, Valdés B, Contreras V, Colombo J, Almazán J. Effects of the ductility capacity on the seismic performance of cross-laminated timber structures equipped with frictional isolators. Buildings. 2025;15(8):1208. doi: 10.3390/buildings15081208 [DOI] [Google Scholar]
  • 28.Wenzel A, Vera S, Guindos P. Energy and structural optimization of mid-rise light-frame timber buildings for different climates and seismic zones in Chile. Eur J Wood Prod. 2024;82(4):967–82. doi: 10.1007/s00107-024-02085-z [DOI] [Google Scholar]
  • 29.Luo J, Guo Y, Liu Z, Hu Q, Hoque MA, Ahmed A. Enhancing deep line segment detection and performance evaluation for wood: a deep learning approach with experiment-based, domain-specific implementations. Forests. 2024;15(8):1393. doi: 10.3390/f15081393 [DOI] [Google Scholar]
  • 30.Luk SH. Machine learning-based methods for the seismic damage classification of RC buildings. Buildings. 2025;15(14):2395. doi: 10.3390/buildings15142395 [DOI] [Google Scholar]

Decision Letter 0

Dajiang Geng

9 Dec 2025

Dear Dr. Zhang,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jan 23 2026 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols ..

We look forward to receiving your revised manuscript.

Kind regards,

Dajiang Geng

Academic Editor

PLOS One

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS One has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, we expect all author-generated code to be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. We note that your Data Availability Statement is currently as follows: All relevant data are within the manuscript and its Supporting Information files.

Please confirm at this time whether or not your submission contains all raw data required to replicate the results of your study. Authors must share the “minimal data set” for their submission. PLOS defines the minimal data set to consist of the data required to replicate all study findings reported in the article, as well as related metadata and methods (https://journals.plos.org/plosone/s/data-availability#loc-minimal-data-set-definition).

For example, authors should submit the following data:

- The values behind the means, standard deviations and other measures reported;

- The values used to build graphs;

- The points extracted from images for analysis.

Authors do not need to submit their entire data set if only a portion of the data was used in the reported study.

If your submission does not contain these data, please either upload them as Supporting Information files or deposit them to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of recommended repositories, please see https://journals.plos.org/plosone/s/recommended-repositories.

If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially sensitive information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent. If data are owned by a third party, please indicate how others may request data access.

4. If the reviewer comments include a recommendation to cite specific previously published works, please review and evaluate these publications to determine whether they are relevant and should be cited. There is no requirement to cite these works unless the editor has indicated otherwise.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously? -->?>

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available??>

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.-->

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English??>

Reviewer #1: Yes

Reviewer #2: Yes

**********

Reviewer #1: This paper aims to enhance the seismic safety performance of wooden structures by establishing the GBRF-SCO framework using artificial intelligence to predict and optimize seismic performance parameters. The research results indicate that it is an effective AI-based method for seismic optimization of wood components. However, there are still many deficiencies in the article that need to be further revised before considering publication.

1. Title and Abstract

(1) Check the spelling of "aptimization" in the title to ensure its professionalism and accuracy. At the same time, the affiliation information of the corresponding author "Nanning Normal University, Nanning, Jiangsu" contains a geographical error. Please correct it before submission.

(2) The abstract mentions that the dataset contains 4,000 samples of different configurations of wooden structures, but it does not specify the source distribution of the samples (such as whether they cover different regions and different eras of wooden structures). It is recommended to supplement this information.

(3) The abstract only emphasizes that the model accuracy rate is 0.949, but does not clearly state the specific scenario corresponding to this accuracy rate (such as whether it is the prediction accuracy rate for inter-story drift or roof drift). It is recommended to make a clear distinction.

2. Literature Review and Research Gap

(4) The literature review section summarizes the application of existing AI in the seismic optimization of wooden structures, but does not highlight the essential differences between the GBRF-SCO framework proposed in this study and the existing ones in terms of core technical routes (such as hyperparameter optimization methods and model fusion logic). It is recommended to supplement the comparative analysis to strengthen the innovative positioning of this study.

3. Research Methods and Data Processing

(5) The data preprocessing section mentions the use of Robust Scaling and Isolation Forest for normalization and outlier detection, but does not explain the criteria for determining outliers (such as the setting of the contamination parameter in Isolation Forest) and the handling methods (such as deletion or replacement). It is recommended to supplement these details to ensure the reproducibility of data processing.

(6) In the GBRF-SCO framework, the specific process of SCO optimizing the hyperparameters of GBRF (such as the basis for selecting population size and number of iterations) is not detailed. It is recommended to supplement the rationality argument of hyperparameter settings to enhance the scientific nature of the method.

(7) In Table 3, the units of parameters such as "Modulus_of_Elasticity" in the sample data are not labeled. It is recommended to supplement the standard units of all physical parameters to ensure the standardization of data representation.

4. Experimental Results and Analysis

(8) In the confusion matrix analysis of Section 4.8, the classification criteria for the three categories of roof displacement ("low, medium, high") are not explained (such as the basis for setting the threshold). It is recommended to clarify the classification rules to ensure the consistency of result interpretation.

(9) In Section 4.9, it is not specified whether the hyperparameters of the comparison models (such as SVM, XGBoost) have been optimized. If not, it may lead to unfair comparison results. It is recommended to supplement the parameter configuration of the comparison models or explain that the default optimal parameters are used to ensure the objectivity of the comparison.

5. Discussion and Conclusion

(10) Section 5.1 "Limitation and future work" only mentions the limitation of data sources. It is recommended to further discuss the possible technical limitations of this method, such as the sensitivity of the model to the quality/ completeness of input parameters, computational cost, and extrapolation ability in extreme earthquakes or unconventional types of wooden structures.

Reviewer #2: 1.It is recommended that the author specify in the abstract which seismic performance parameters were optimized in the design, providing more detailed information.

2.The Introduction should be restructured to provide a more detailed and systematic review of prior work. The current organization does not meet the standard expected for a literature review, hindering the reader's understanding of the current research developments in this area.

3.What is the intended purpose of the literature review here? How does it differ from providing a list of citations?

4.The literature review should focus on timber structure seismic performance. Please explain why you discussed the steel-grout connection.

5.What is the bar chart in Figure 2 meant to represent? The current description does not clearly communicate the intended message.

6.Please correct the numbering format for the equations to align with standard academic conventions.

7.Please clarify how the model's credibility was established. Specifically, was it validated or compared against any existing work before the analysis in “4.Performance analysis and discussion”?

8.It is recommended that the authors streamline the Conclusions section to make it more focused and impactful. The current version is somewhat lengthy; condensing it would help to highlight the core findings and main message of the study more effectively.

9.It is noted that your manuscript needs careful editing by someone with expertise in technical English editing, paying particular attention to English grammar, sentence structure, and the appropriate use of voice, so that the goals and results of the study are clear to the reader.

**********

what does this mean? ). If published, this will include your full peer review and any attached files.). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our For information about this choice, including consent withdrawal, please see our Privacy Policy .-->

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

To ensure your figures meet our technical requirements, please review our figure guidelines: https://journals.plos.org/plosone/s/figures

You may also use PLOS’s free figure tool, NAAS, to help you prepare publication quality figures: https://journals.plos.org/plosone/s/figures#loc-tools-for-figure-preparation.

NAAS will assess whether your figures meet our technical requirements by comparing each figure against our figure specifications.

PLoS One. 2026 Apr 8;21(4):e0341961. doi: 10.1371/journal.pone.0341961.r002

Author response to Decision Letter 1


4 Jan 2026

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

Author Response:

We thank the reviewer for the reminder regarding PLOS ONE formatting requirements. The manuscript has been fully revised in accordance with the official PLOS ONE style templates, including correct file naming, text layout, headings, tables, figures, and reference formatting. All alignments and stylistic elements have been carefully updated to ensure full compliance with the journal’s formatting guidelines.

2. Please note that PLOS One has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, we expect all author-generated code to be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

Author Response: We have revised the pseudocode to more accurately reflect the proposed GBRF–SCO framework, clarifying the model flow, optimization strategy, and output formulation. The updated version now aligns with the regression-based seismic performance objectives, clearly defined fitness functions, and hyperparameter tuning process. This revision improves methodological clarity, consistency with the manuscript, and reproducibility.

Pseudo Code 1: GBRF–SCO Framework for Seismic Performance Prediction in Timber Structures

Input:

X=[x1,x2,…,xm]   # Structural and seismic features

Y=[y1,y2,…,ym]   # Continuous seismic response (roof displacement / drift)

P              # SCO population size

S              # SCO iterations

LB,UB           # Hyperparameter bounds

Output:

Yfinal      # Optimized seismic response prediction

Step 1: Data Preprocessing

1.Remove outliers using Isolation Forest

2.Normalize features using Robust Scaling

3.Split data into training and testing sets

Step 2: Gradient Boosted Random Forest (GBRF) Modeling

1.Initialize GBRF hyperparameters:

θ={nestimators,η,dmax)

2.Train GBRF model on XtrainYtrain

3.Predict seismic response Y

Step 3: Scalable Cheetah Optimizer (SCO)

1.Initialize cheetah population:

Wj0=LB+rand(0,1)×(UB−LB)

2.For s=1to S:

Evaluate fitness:

Fitness(Wjs)=RMSE(Y,Y)

Identify best cheetah W∗

Update position:

Wjs+1=σγ(k+v)−Wjs+αs(W∗−Wjs)

Apply boundary constraints

3.Return optimal hyperparameters θ∗

Step 4: Final Prediction

1.Retrain GBRF using θ∗

2.Generate optimized seismic response: Yfinal

Step 5: Seismic Damage Classification (Optional)

Class=Low,Yfinal<Q33Medium,Q33≤Yfinal<Q66High,Yfinal≥Q66

Return: Yfinal

3. We note that your Data Availability Statement is currently as follows: All relevant data are within the manuscript and its Supporting Information files.

Please confirm at this time whether or not your submission contains all raw data required to replicate the results of your study. Authors must share the “minimal data set” for their submission. PLOS defines the minimal data set to consist of the data required to replicate all study findings reported in the article, as well as related metadata and methods (https://journals.plos.org/plosone/s/data-availability#loc-minimal-data-set-definition).

For example, authors should submit the following data:

- The values behind the means, standard deviations and other measures reported;

- The values used to build graphs;

- The points extracted from images for analysis.

Authors do not need to submit their entire data set if only a portion of the data was used in the reported study.

If your submission does not contain these data, please either upload them as Supporting Information files or deposit them to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of recommended repositories, please see https://journals.plos.org/plosone/s/recommended-repositories.

If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially sensitive information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent. If data are owned by a third party, please indicate how others may request data access.

Author Response: The data used in this study are publicly available from the Kaggle repository. The dataset can be accessed and downloaded free of charge from Kaggle at the corresponding dataset page. No restrictions apply to the availability of these data.

Ethical statement

The Timber Seismic Performance Dataset referenced in this work is a secondary dataset publicly available through the Kaggle repository (https://www.kaggle.com/datasets/freshersstaff/timber-seismic-performance-dataset/data). This dataset has been used under the terms and conditions provided by Kaggle. The data are accessible without restriction to any researcher, and no identifiable human subjects or sensitive personal information are included in the dataset.

4. If the reviewer comments include a recommendation to cite specific previously published works, please review and evaluate these publications to determine whether they are relevant and should be cited. There is no requirement to cite these works unless the editor has indicated otherwise.

Reviewer #1: This paper aims to enhance the seismic safety performance of wooden structures by establishing the GBRF-SCO framework using artificial intelligence to predict and optimize seismic performance parameters. The research results indicate that it is an effective AI-based method for seismic optimization of wood components. However, there are still many deficiencies in the article that need to be further revised before considering publication.

1.Title and Abstract

(1)Check the spelling of "aptimization" in the title to ensure its professionalism and accuracy. At the same time, the affiliation information of the corresponding author "Nanning Normal University, Nanning, Jiangsu" contains a geographical error. Please correct it before submission.

Author Response: Thank you for pointing out these issues. The spelling error in the title has been corrected from “aptimization” to “optimization.” This corrections have been updated in the revised manuscript.

(2)The abstract mentions that the dataset contains 4,000 samples of different configurations of wooden structures, but it does not specify the source distribution of the samples (such as whether they cover different regions and different eras of wooden structures). It is recommended to supplement this information.

Author Response: We appreciate this valuable suggestion. The abstract has been revised to clarify that “The dataset consists of 4,000 timber building samples obtained from a publicly available Kaggle repository (Timber Seismic Performance Dataset).” The dataset covers diverse structural configurations, material properties, and seismic intensity scenarios generated through simulation-based analysis, rather than being restricted to a specific geographical region or historical period. This clarification has been added to enhance transparency regarding the data source and provenance.

(3)The abstract only emphasizes that the model accuracy rate is 0.949, but does not clearly state the specific scenario corresponding to this accuracy rate (such as whether it is the prediction accuracy rate for inter-story drift or roof drift). It is recommended to make a clear distinction.

Author Response: Thank you for this comment. The abstract has been revised to explicitly state that the reported “accuracy of 0.949, which corresponds to the classification of roof displacement levels (low, medium, high) under seismic loading conditions.” This distinction ensures clarity regarding the specific prediction task and performance scenario evaluated in the study.

2.Literature Review and Research Gap

(4)The literature review section summarizes the application of existing AI in the seismic optimization of wooden structures, but does not highlight the essential differences between the GBRF-SCO framework proposed in this study and the existing ones in terms of core technical routes (such as hyperparameter optimization methods and model fusion logic). It is recommended to supplement the comparative analysis to strengthen the innovative positioning of this study.

Author Response: We thank the reviewer for this constructive suggestion. A new subsection titled “Distinction Between Existing AI-Based Approaches and the Proposed GBRF-SCO Framework” has been added to the Literature Review. This section clearly contrasts the proposed framework with existing methods in terms of hyperparameter optimization strategy, model fusion logic, and optimization–prediction coupling, thereby strengthening the innovative positioning of this study. All additions are marked in red in the revised manuscript.

2.5 Distinction Between Existing AI-Based Approaches and the Proposed GBRF-SCO Framework

Previous AI-based seismic researches on timber structures mainly rely on single learning models with manually tuned or grid-searched hyperparameters, which often suffer from high computational cost and suboptimal convergence in high-dimensional design spaces. In contrast, the proposed GBRF-SCO framework integrates a hybrid Gradient Boosted Random Forest model with the Scalable Cheetah Optimizer for adaptive, population-based hyperparameter tuning. By coupling prediction and optimization within a unified learning loop, the framework enhances nonlinear modeling capability, robustness, and design-oriented seismic performance optimization.

3.Research Methods and Data Processing

(5)The data preprocessing section mentions the use of Robust Scaling and Isolation Forest for normalization and outlier detection, but does not explain the criteria for determining outliers (such as the setting of the contamination parameter in Isolation Forest) and the handling methods (such as deletion or replacement). It is recommended to supplement these details to ensure the reproducibility of data processing.

Author Response: We thank the reviewer for this important comment. The Data Preparation section has been revised to explicitly describe the Isolation Forest contamination setting, outlier decision criteria, and handling strategy. Specifically, the contamination parameter, detection threshold, and the approach used for managing detected outliers are now clearly stated to ensure full reproducibility. All added details are marked in red in the revised manuscript.

Outlier detection: The Isolation Forest algorithm is used to find anomalies in the seismic and structural dataset by looking for samples that are different from the rest. The contamination parameter is set to 0.05, which means that about 5% of the observations are likely to be outliers. Samples with anomaly scores below the decision threshold are considered anomalous and taken out of the dataset because they show physically inconsistent or severe seismic–structural pairings. This method cuts down on noise and bias, makes the data more reliable, and makes sure that the samples kept are a true representation of how timber structures behave in genuine seismic conditions. This makes model training more stable and effective.

(6)In the GBRF-SCO framework, the specific process of SCO optimizing the hyperparameters of GBRF (such as the basis for selecting population size and number of iterations) is not detailed. It is recommended to supplement the rationality argument of hyperparameter settings to enhance the scientific nature of the method.

Author Response: The manuscript has been revised to clarify the rational basis of the SCO hyperparameter settings. The population size (r = 8-10) was selected to maintain sufficient solution diversity while minimizing computational overhead, and the iteration range (S = 15-20) was determined based on convergence stability, beyond which marginal performance gains were negligible.

Answer:

The selection of SCO hyperparameters depends on how well they converge, how fast they can be computed, and how many dimensions the seismic optimization problem has. The population size (r = 8-10) makes sure that there are enough different solutions without making the computation too expensive in feature spaces with a lot of dimensions. Smaller populations limit research, but bigger ones provide slight advances in accuracy. Convergence analysis, which is when the optimization error stabilizes, tells you the maximum number of iterations (S = 15-20). This setup strikes a good mix between exploration and exploitation, which makes GBRF hyperparameter tuning steady and dependable.

(7)In Table 3, the units of parameters such as "Modulus_of_Elasticity" in the sample data are not labeled. It is recommended to supplement the standard units of all physical parameters to ensure the standardization of data representation.

Author Response: Thank you for this observation. Standard units have now been added to all physical parameters in Table 3, including modulus of elasticity, mass, acceleration, and displacement variables, to ensure clarity, consistency, and standardized data representation. The revised table is marked in red in the manuscript.

Answer:

Table 3: sample data from the Timber Seismic Performance Dataset

Parameter Sample 1 Sample 2 Sample 3 Sample 4

Building Height (m) 36.21781 76.55 61.23958 51.90609

Number of Storeys 11 6 7 5

Story Height (m) 3.91272 3.40562 2.6579 3.35902

Wall Thickness (m) 0.16088 0.5326 0.38316 0.45603

Material Density (kg/m³) 701.6804 496.9688 467.4328 557.7363

Modulus of Elasticity (pa) 9.85E+09 1.14E+10 1.37E+10 9.06E+09

Damping Coefficient 0.02285 0.03804 0.01479 0.05187

Damping Ratio per Storey 0.0332 0.02278 0.04835 0.03449

Floor Mass (kg) 74798.29 19779.63 77220.55 29065.37

Natural Frequency (hz) 6.47606 7.36455 4.99899 3.67033

Peak Ground Acceleration (g) 0.24833 0.2076 0.35973 0.2887

Spectral Acceleration (m/s²) 2.70666 2.51013 3.1417 3.13824

Lateral Load Resisting Ratio 0.47871 0.28382 0.37719 0.78881

Inter Storey Drift (m) Timber Pegs Hybrid Connectors Hybrid Connectors Timber Pegs

Roof Displacement (m) Pile Pile Pile Pile

Building Height (m) Hip Hip Gable Hip

Number of Storeys Shear Wall Moment Frame Shear Wall Moment Frame

Story Height (m) Symmetric Asymmetric Symmetric Asymmetric

Wall Thickness (m) 0.00246 0.00602 0.00492 0.00729

Material Density (kg/m³) 0.02655 0.03641 0.03649 0.03525

4.Experimental Results and Analysis

(8)In the confusion matrix analysis of Section 4.8, the classification criteria for the three categories of roof displacement ("low, medium, high") are not explained (such as the basis for setting the threshold). It is recommended to clarify the classification rules to ensure the consistency of result interpretation.

Author Response:

Thank you for the constructive comment. Section 4.8 has been revised to clearly define the classification rules for roof displacement using data-driven quantile-based thresholds (low, medium, high), ensuring ordinal consistency and balanced class distribution. This addition improves the transparency, interpretability, and reproducibility of the confusion matrix analysis.

Answer: Based on quantile thresholds from the dataset's empirical distribution, roof displacement values are divided into three ordinal classifications such as low, medium, and high. Values of displacement that are less than the 33rd percentile are considered low, values that are between the 33rd and 66th percentiles are considered me

Attachment

Submitted filename: Review comments.docx

pone.0341961.s002.docx (14KB, docx)

Decision Letter 1

Dajiang Geng

15 Jan 2026

<p>AI-assisted aptimization design of seismic performance parameters for timber structures

PONE-D-25-59376R1

Dear Dr. Xuan Zhang,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager®  and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact  and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact billing support ..

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Dajiang Geng

Academic Editor

PLOS One

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions??>

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously? -->?>

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available??>

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.-->

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English??>

Reviewer #1: Yes

Reviewer #2: Yes

**********

Reviewer #1: (No Response)

Reviewer #2: (No Response)

**********

what does this mean? ). If published, this will include your full peer review and any attached files.). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our For information about this choice, including consent withdrawal, please see our Privacy Policy .-->

Reviewer #1: No

Reviewer #2: No

**********

Attachment

Submitted filename: renamed_fedfb.docx

pone.0341961.s003.docx (13.5KB, docx)

Acceptance letter

Dajiang Geng

PONE-D-25-59376R1

PLOS One

Dear Dr. Zhang,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS One. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

You will receive an invoice from PLOS for your publication fee after your manuscript has reached the completed accept phase. If you receive an email requesting payment before acceptance or for any other service, this may be a phishing scheme. Learn how to identify phishing emails and protect your accounts at https://explore.plos.org/phishing.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Dajiang Geng

Academic Editor

PLOS One

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. We have uploaded the minimal dataset underlying the findings of this study as Supporting Information with the submission.

    In addition, the dataset is publicly available without restriction from the Kaggle repository at the following link: https://www.kaggle.com/datasets/freshersstaff/timber-seismic-performance-dataset/data

    (XLSX)

    pone.0341961.s001.xlsx (665KB, xlsx)
    Attachment

    Submitted filename: Review comments.docx

    pone.0341961.s002.docx (14KB, docx)
    Attachment

    Submitted filename: renamed_fedfb.docx

    pone.0341961.s003.docx (13.5KB, docx)

    Data Availability Statement

    All relevant data are within the manuscript.


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES