Abstract
Understanding air quality requires a comprehensive understanding of its various factors. Most of the association rule techniques focuses on high frequency terms, ignoring the potential importance of low- frequency terms and causing unnecessary storage space waste. Therefore, a dynamic genetic association rule mining algorithm is proposed in this paper, which combines the improved dynamic genetic algorithm with the association rule mining algorithm to realize the importance mining of low- frequency terms. Firstly, in the chromosome coding phase of genetic algorithm, an innovative multi-information coding strategy is proposed, which selectively stores similar values of different levels in one storage unit. It avoids storing all the values at once and facilitates efficient mining of valid rules later. Secondly, by weighting the evaluation indicators such as support, confidence and promotion in association rule mining, a new evaluation index is formed, avoiding the need to set a minimum threshold for high-interest rules. Finally, in order to improve the mining performance of the rules, the dynamic crossover rate and mutation rate are set to improve the search efficiency of the algorithm. In the experimental stage, this paper adopts the 2016 annual air quality data set of Beijing to verify the effectiveness of the unit point multi-information coding strategy in reducing the rule storage air, the effectiveness of mining the rules formed by the low frequency item set, and the effectiveness of combining the rule mining algorithm with the swarm intelligence optimization algorithm in terms of search time and convergence. In the experimental stage, this paper adopts the 2016 annual air quality data set of Beijing to verify the effectiveness of the above three aspects. The unit point multi-information coding strategy reduced the rule space storage consumption by 50%, the new evaluation index can mine more interesting rules whose interest level can be up to 90%, while mining the rules formed by the lower frequency terms, and in terms of search time, we reduced it about 20% compared with some meta-heuristic algorithms, while improving convergence.
Introduction
In recent years, environmental pollution has emerged as a critical concern, with air pollution gaining increasing prominence. Pollution arises from numerous sources, including non-compliant factory emissions, vehicle exhaust emission, construction-related dust, and agricultural practices such as straw burning. These unscientific activities lead to abnormal changes in the concentration of carbon, nitrogen, sulfur, particulate matter with diameters less than 2.5 micrometers (PM2.5) and 10 micrometers (PM10), as well as O3 in the air, resulting in atmospheric pollution. Air Quality Index (AQI) serves as a metric for ranking air pollution levels, influenced by a myriad of factors. These encompass meteorological elements (cloud cover, sunlight, precipitation, wind speed and direction) and geographical variables (altitude, latitude, and longitude). Previous studies have typically associated AQI categories with individual influencing factors, determining whether the category is positively, negatively, or not correlated with the given factor. This study aims to uncover the correlation between AQI and multiple factors within a specific category, as well as the degree of this correlation. Furthermore, it seeks to explore the correlation between certain factors under the sample data. The derived rules can serve as guiding principles for air quality improvement. For instance, mining from the sample data reveals that the air quality within the second category, denoted as ’good’, is influenced by the concentration of SO2 in the second gradient and NO2 concentration within the first gradient interval range.
Association rule mining (ARM) [1] is an unsupervised learning technique in the field of data mining, which is used to mine rules from a specific scenario, research or transaction process database. Its original purpose was to provide interesting relationships, associations, or frequent patterns between sets of items in database transactions. Intriguingly, ARM’s genesis can be traced back to an observed co-purchasing trend of beer and diapers in the renowned Walmart supermarket chain. Today, ARM finds applications across diverse sectors. In retail, it informs product placement strategies to optimize promotions; in medicine, it helps uncover relationships between ailments and treatment strategies. For example, Elif et al. used numerical ARM to identify potentially important rules between Parkinson’s disease and voice change characteristics [2, 3]; in recommendation engines, it facilitates matching users based on shared preferences; and in safety research, ARM has been instrumental in identifying principal causative factors behind accidents [4–6]. In mechanical fault diagnostics, ARM facilitates rapid fault source identification, enabling immediate remediation [7]. Considering the above broad application scenarios, this study employs ARM to elucidate correlations among influencing factors in air quality, particularly within the Beijing region of China. However, several challenges need to be overcome before successful application in finding associations between air quality factors:
·Traditional ARM is carried out on the basis of discrete data types. Since some influencing factors of air quality are continuous, it is difficult to mine rules in the case of pre-processing and integration of different types of data.
·Most of the previous ARM techniques were designed to filter out items with a certain frequency by artificially setting the support degree, and then explore their relationships, which made it challenging to find the relationship between items with low frequency. Such an operation can undervalue the significance of anomalies in air quality datasets, resulting in biased interpretations of factors impacting air quality.
·The performance of ARM combined with other algorithms still has room for improvement. For example, in the aspect of space storage, the traditional technologies, including the Apriori algorithm, demand multiple database traversals when discerning item set associations and store each item contained in the database transaction with one unit storage space, which yield candidate item sets at exponential magnitudes and wastes storage space. In the mining of rules, the quality and time efficiency of rules can still be improved.
Contribution of this paper
To solve these problems, a novel genetic algorithm-Dynamic Genetic Algorithm (DGA) is introduced in this paper. The key contributions of this study are as follows:
The manuscript proposes the concept of dynamic threshold of mutation rate and crossover rate, which improves its ability to accurately locate the optimal solution, thus minimizing the convergence time.
A well-conceived encoding strategy aligns seamlessly with the research objectives, as reflected in the efficient complexity of the Dynamic Genetic Association Rule Mining (DGAARM) algorithm.
The weighted optimization of the overall evaluation index can ensure that more valuable rules can be extracted.
Related work
Agrawal et al. [8] first introduced the concept of ARM which provides an opportunity to discover item-to-item relationships from a data set containing a large number of variables [9]. Subsequently, Apriori algorithm [10] was proposed and received wide attention. Despite its broad attention, the Apriori algorithm, in its quest for frequent itemsets, produces an excessive number of candidate items. This led to innovations like the approach by Han et al. [11], who employed a tree data structure to organize transaction data and conducted depth-first traversal, resulting in the renowned Frequent Pattern Growth (FP-Growth) algorithm. However, this requires preliminary tasks such as item frequency ranking and the laborious construction of a conditional Frequent Pattern Tree (FP-tree). The emergence of Eclat algorithm changes the horizontal representation of data to vertical representation, and turns multiple scanning of database for Apriori and FP-Growth to only two times when calculating item set support, which is completed by intersecting Tidset after vertical representation of data. However, because Eclat algorithm takes a long time to find the intersection when there is a large amount of data, Zhang et al. [12] used minwise hashing and estimator to quickly calculate the intersection size of multiple item sets, thus improving its efficiency. Based on the fact that Eclat consumes a large amount of memory space and computation time in data reading and computation under a large amount of data, and generates a large amount of redundant data in the process, Wang et al. [13] improved the pruning strategy of Eclat, which reduced the generation of redundant frequent items, and improved the efficiency of the algorithm. In recent years, many researchers have improved on the basis of the above two algorithms. For example, in the analysis of tower crane accidents, Liu et al. [14] introduced the interest degree (I) model with upper and lower bound idea, lifting degree and leverage ratio evaluation indexes based on Apriori algorithm, which reduced the number of redundant rules, but did not improve the algorithm performance. Given the large memory occupied by the FP-Growth algorithm to construct the pattern tree by using the entire transaction database, the low operation efficiency of the algorithm, and the poor timeliness of data mining, Yu and Liu et al. [15] proposed the MFP-tree algorithm. When traversing the transaction database for the first time, the algorithm calculates the support degree of all items, deletes the items that do not meet the threshold according to the artificially set support threshold, re-sorts the items of each transaction, constructs a database subset of each item according to the frequent 1-item set, and then carries on the FP-Growth algorithm on this subset. Experiments show that MFP-tree algorithm has certain advantages over FP-Growth algorithm when mining database is larger or constraint conditions are strict.
Furthermore, many ARM applications rely heavily on support and confidence degree thresholds during extraction mining. For instance, Wang et al. [16] applied a support degree of 0.2 using the MapReduce model to enhance the Apriori algorithm, while Liu et al. [17] utilized a confidence threshold of 0.5 with the parallel FP-Growth algorithm to decipher rules between temperature and salinity in marine Argo datasets. This methodology can inadvertently filter out items and rules below set thresholds, however sometimes infrequent items and rules are of interest to researchers instead.
The meta-heuristic algorithm based on the improvement of heuristic algorithm includes Genetic Algorithm (GA) and swarm intelligence algorithms such as Particle Swarm Optimization (PSO) algorithm and whale optimization algorithm (WOA). Among them, GA encompasses many variants such as classical, parallel, hierarchical, adaptive, and hybrid algorithms [18–21], and showed good performances for optimization problems. Since ARM is widely used in various fields for knowledge discovery or pattern association, some researchers combine heuristic algorithms with association rules to improve the time performance and result optimization of algorithms. For example, S. Sharmila et al. [22] combines WOA with fuzzy logic to identify frequent items and generate association rules. In the study of numerical association rules, Elif et al. [23] proposes a new hybrid multi-objective evolutionary optimization algorithm based on differential evolution (DE) and sine and cosine algorithm. The sine-cosine algorithm can effectively prevent premature convergence and stagnation in the iterative process, and improve the overall search ability and convergence performance of the algorithm. Given ARM only considers the frequency of items in the item set to find the item set of interest, which cannot reflect the usefulness or preference of users to quantify products with different values, Kannimuthu et al. [24] introduced a high-utility itemset mining algorithm. Adopting GA to optimize the PSO algorithm to avoid the combination explosion problem and the problem of early stagnation of algorithm search, it turned out that the number of candidate item sets is reduced effectively and the convergence performance of the algorithm is improved. In order to avoid the combination explosion problem in the study of web service composition, S. Kannimuthu et al. [25] proposed a hybrid genetic algorithm (HGA), which combines quantum operators and classical genetic operators, to mine efficient web service composition. The chromosome constructed by superposition qubits based on quantum computing model achieves good results in terms of running time and memory consumption. In addition, relevant researchers regard the support, confidence and other evaluation indicators in ARM as multiple objectives and adopt multi-objective optimization association rule mining. For example, Tyagi et al. [26] extracted valuable rules by multi-objective particle swarm optimization (MOPSO) in the collaborative filtering of recommendation system to improve the recommendation quality. In addition, since users have prior knowledge and research trends of some key items in practical applications, association rules containing key items are more valuable and meaningful for these users. Therefore, Hu et al. [27] proposed the Animal Dynamic Migration Optimization (ADMO) algorithm for directional mining rules. By changing the constant direction migration of animals in the original animal migration algorithm to the dynamic direction correction mode, good results are obtained in key rules, rule optimization, memory consumption and execution time.
AQI is a gauge of daily air quality, segmented into six categories from Class I to Class VI. Each AQI level has distinct implications for human health, influencing t well-being and societal progress. Initiatives to understand the determinants of air quality, diminish pollution sources, and thwart the interplay of multiple pollutants are pivotal for air quality enhancement. Current research has made significant strides in deciphering the factors influencing air quality. For example, Li et al. [28] undertook linear correlation and multiple regression analyses on monthly air quality variations and meteorological elements across cities. The meta-analysis based on correlation and regression coefficients showed the relationship between certain pollution factors and meteorological variables. Notably, PM2.5 concentrations showed correlation with all meteorological metrics, except wind speed. In contrast, PM10 and O3 concentrations exhibited links with all meteorological variables; however, O3’s correlation direction with meteorological indicators deviated from that of PM2.5 and PM10. Zhu [29] delved into the spatiotemporal and socio-economic attributes of regional air pollution by devising a panel data gray correlation clustering model and a gray entropy test model. Duan et al. [30] used GA to optimize subregion-level priority of precursor emission reductions and combined Self-Organizing Map (SOM) and WRF-CAMx for the collaborative control of PM2.5 and O3 in Beijing-Tianjin-Hebei and the surrounding area (BTHSA, “2 + 26” cities).
Organization of this paper
The structure of this paper is as follows: The Materials and methods section introduces the general framework, data pre-processing of the air quality data set used, the basic concept of ARM and DGAARM algorithm. The Result and discussion section introduces the experimental results and related discussions. Finally, a summary of the work in the Conclusions section of this paper is presented.
Materials and methods
General framework
In this study, we apply the DGAARM algorithm to optimize the performance of ARM. We realize the mining of interesting air quality association rules without artificially setting minimum support threshold, and design a unique coding method to optimize the spatial storage of the rules, and optimize the convergence performance by combining dynamic crossover and variation rate in the mining process. Initially, we pre-process the 2016 annual air data from Beijing, which involves data extraction, transformation, and loading. Subsequently, we target and code the chromosome genes in accordance with the problem’s specificity. We then employ the DGAARM algorithm to unearth the rules governing air quality influencing factors, before comparing its performance with other classical association rule algorithms. The overarching framework of the proposed method is depicted in Fig 1.
Fig 1. General framework for mining association rules of air quality impact factors based on DGAARM.
Data pre-processing
The data pre-processing stage is divided into three parts: firstly, Data Extraction, then, Data Transformation, lastly, Data Loading, as shown in the left part of Fig 1. In the data extraction phase, the experimental dataset is utilized in this study which was obtained from the environmental cloud of Nanjing Yunchuang Big Data Technology Co., LTD. (Nanjing, China). We accessed hourly meteorological records data and hourly air quality monitoring data for Beijing, spanning from 1 January 2016 to 31 December 2016, which were recorded by 12 monitoring sites. We spliced two parts of the data and selected one of the sites with a total of 8784 records. We loaded these data into the Pandas library’s DataFrame object for feature extraction. The properties of the raw data are detailed in Table 1.
Table 1. The raw data description table.
| Statistical name | Statistical parameters |
|---|---|
| Traits | TIME, CITY CODE, Weather conditions, Temperature, Body temperature, Barometric pressure, Relative humidity, Rainfall, Wind direction, Wind speed, Air quality index, PM2.5 concentration, PM10 concentration, NO2 concentration, SO2 concentration, O3 concentration, CO concentration |
| Number of features | 18 |
| Number of data bars | 8784 |
Then irrelevant attributes such as time, city-specific invariant attributes (given the city is consistently Beijing), and body temperature were removed. Subsequently, the remaining features were numerically assessed, leading to the deletion of non-numeric data types. The results of the post-extraction are detailed in Table 2. In the data transformation phase, the data were discretized, and the results are shown in Table 3. Weather conditions were manually discretized, while the final seven features were categorized based on the Ambient Air Quality Standard (GB3095-2012) and AQI Technical Provisions (for Trial Implementation) HJ633-2012. The dataset contains weather categories such as "sunny" and "haze", among 16 other classification categories. Temperature data, ranging from -15.1°C to 37.3°C, was organized into 17 classes including descriptors like "deep chill" and "the Great Cold". Air pressure is classified using neighboring relative air pressure values: values below one standard atmosphere are deemed "low pressure" and those above as "high pressure". Relative humidity classifications are "dry" for values below 30%, "humid" for above 80%, and "normal" for the intermediate range. Rainfall data, with a range from 0mm to 32.5mm, is categorized into four levels: R0 to R3. These classifications stem directly from the specific range of each feature’s data. Finally, load the final data into the DGAARM algorithm model, completing the classical ETL (Extract, Transform, Load) processes of data loading, feature extraction, and data transformation.
Table 2. Data description table after ETL operation.
| Statistical name | Statistical parameters |
|---|---|
| Traits | Weather conditions, Temperature, Barometric pressure, Relative humidity, Rainfall, Wind direction, Wind speed, Air quality index, PM2.5 concentration, PM10 concentration, NO2 concentration, SO2 concentration, O3 concentration, CO concentration |
| Number of features | 14 |
| Number of data bars | 6430 |
Table 3. Characteristics description table after discretizing operation.
| Feature Name | Values after discretizing operation |
|---|---|
| Weather conditions | Sunny, Hazy, Cloudy, Overcast, Light rain, Moderate to heavy rain, Heavy rain, Rain showers, Thunderstorms, Fog, Sleet, Light snow, Moderate to heavy snow, Floating layer, Moderate rain, Rainstorm. |
| Temperature | Deep Chill (-15~20°C), The Great Cold (-10~15°C), Little cold (-10~-5°C), Lightly cold (-5~0°C), Slightly cold (0~5°C), Cool (5~10°C), Warm and cool(10~12°C), Slightly warm and cool (12~14°C), Mild (14~16°C), Slightly mild (16~18°C), Warm (18~20°C), Warm but not hot (20~22°C), Slightly hot (22~25°C), Hot (25~28°C), Summer hot (28~30°C), Very hot (30~35°C), Extremely hot (35~39°C). |
| Barometric pressure | Low-pressure, High pressure. |
| Relative humidity | Dry (<30%), Normal (30%~80%), Humidity (>80%). |
| Rainfall | R0 (< = 5mm), R1 (5~15mm), R2 (15~30mm), R3 (30~70mm). |
| Wind direction | East, Southeast, South, Southwest, West, Northwest, North, Northeast, No sustained wind direction. |
| Wind power | Light breeze, Force3 wind, Force4 wind, Force5 wind. |
| Air quality index | AQI-1 ~ AQI-6a |
| PM2.5 concentration | PM2_5_I ~ PM2_5_VIIb |
| PM10 concentration | PM10_I ~ PM10_VIIIc |
| NO2 concentration | NO2_I ~ NO2_IVd |
| SO2 concentration | SO2_I ~ SO2_IIIe |
| O3 concentration | O3_I ~ O3_Vf |
| CO concentration | CO_I ~ CO_IIIg |
aAQI has been classified as AQI-I to AQI-VI, representing the first rank to the sixth rank of air quality rank which ranges from 0 to 50, 51 to 100, 101 to 150, 151 to 200, 201 to 300, and more than 300, respectively.
bPM2.5 has been classified as PM2_5_I to PM2_5_VII, which range from 0 to 35μg/m3, 35 to 75μg/m3, 75 to 115μg/m3, 115 to 150μg/m3, and 150 to 250μg/m3, respectively.
cPM10 has been classified as PM10_I to PM10_VIII, which range from 0 to 50μg/m3, 50 to 150μg/m3, 150 to 250μg/m3, 250 to 350μg/m3, 350 to 420μg/m3, 420 to 500μg/m3, 500 to 600μg/m3and more than 600μg/m3, respectively.
dNO2 has been classified as NO2_I to NO2_V, which range from 0 to 40μg/m3, 40 to 80μg/m3, 80 to 180μg/m3, 180 to 280μg/m3and more than 280μg/m3, respectively.
eSO2 has been classified as SO2_I to SO2_III, which range from 0 to 50μg/m3, 50 to 150μg/m3 and more than 150μg/m3, respectively.
fO3 has been classified as O3_I to O3_V, which range from 0 to 100μg/m3, 100 to 160μg/m3, 160 to 2150μg/m3, 215 to 265μg/m3 and more than 265μg/m3, respectively.
gCO has been classified as CO_I to CO_III, which range from 0 to 2 mg/m3, 2 to 4 mg/m3and more than 4 mg/m3, respectively.
ARM
The algorithm for ARM is primarily concerned with identifying patterns in the form of X = >Y within a database, where X and Y are mutually exclusive sets. The process of ARM is bifurcated into two stages: the extraction of frequent item sets and the subsequent discovery of association rules. The initial stage is characterized by the use of support (sup), while the latter stage employs evaluation metrics such as confidence (conf) and lift (lift). Both support and confidence serve as indicators of the robustness of the association rules [31].
Definition 1. Association rules
Consider a transaction database D, where each distinct attribute is represented as a unique item i. This results in an itemset , where n1 signifies the total number of attributes in the database. Let represent the set of transactions, with n2 indicating the overall count of transactions within the database. The association rule takes the form , where , . The left-hand side of the symbol = > is commonly known as the antecedent, while the right-hand side is referred to as the consequent.
Definition 2. Support [32]
Support refers to the proportion of transactions that contain a specific itemset, as determined by ∂(X), relative to the total number of transactions within the database. This is computed using Eq (1).
| (1) |
Definition 3. Confidence [33]
Confidence serves as a metric quantifying the strength of association between the antecedent and the consequent of a rule. A higher confidence value signifies a stronger association between the antecedent and the consequent. It is computed using Eq (2).
| (2) |
Wherein Y is also a subset of the item set I. In this context, X represents the antecedent and Y signifies the consequent, with X∪Y denoting the set encompassing all items of the rule. The confidence of the rule is determined by calculating the ratio of the number of transactions in the database that include all items of the rule to the number of transactions that contain all items in the antecedent of the rule. Consequently, this ratio represents the probability of Y’s occurrence given the occurrence of X.
Definition 4. Lift
Lift serves as an indicator of the extent to which the presence of one item influences the likelihood of another item’s occurrence. It provides insight into the correlation between items, whether positive, negative, or non-existent. It is computed using Eq (3).
| (3) |
A positive correlation exists between X and Y when lift(X→Y)>1, while a negative correlation is observed when lift(X→Y)<1. In instances where lift(X→Y)≡1, X and Y are deemed to be independent, indicating no correlation.
DGAARM algorithm
The DGAARM algorithm proposed in this paper integrates genetic algorithm into ARM to quickly reveal the rule between various air quality factors and Air quality index (AQI) in a specific environment. DGAARM algorithm consists of four key parts: chromosome gene coding, chromosome population initialization, selection during algorithm execution, crossover and mutation operators design, and chromosome population renewal iteration process. By introducing multi-information unit points, dynamic crossover rate and dynamic mutation rate, the optimal solution discovery ability of genetic algorithm is enhanced.
Coding design of genes
The encoding phase of DGAARM focuses on representing association rules in binary codes. Two common methods are the Pittsburgh method, which uses a single chromosome to describe ’n’ association rules, and the Michigan method, which uses one chromosome for each association rule. In this study, the latter approach was employed for chromosome design.
The number of loci in a chromosome is dictated by the transaction database features. In this experiment, 14 features resulted in 14 loci, each encapsulated by the ATGC class, subdivided into the former, center, and latter data domains. The former domain stores a specific discrete category under a feature; the center domain indicates the presence (1) or absence (0) of the item set stored by the former domain in the rule; the latter domain indicates whether the item set is in the predecessor (0) or the posterior (1) of the rule. A representation of the gene locus for air quality characteristics and the rule encoding of is provided in Figs 2 and 3, respectively.
Fig 2. Gene loci under the feature of AQI.
Fig 3. The generated rule case diagram.
Initialization of primordial chromosome population
Each chromosome in the population carries a wealth of information within its genes, marked by random numbers assigned to each domain of every locus. The range of these random numbers varies across domains; in the former domain, the range corresponds to the discrete category count under the feature of that locus, whereas in the center and latter domains, the range is confined to the set {0,1}.
After encapsulating the former, center, and latter domains into a gene, the validity of the newly generated chromosome is verified using the JudgeGene function. This process scrutinizes whether the front and rear sections of the chromosome are populated and checks the chromosome’s existence within the dataset. Should these conditions not be met, the chromosome is regenerated.
The specific algorithmic process can be outlined as follows and is illustrated in Fig 4:
Fig 4. Initialization process of primordial chromosome population.
For each chromosome in the population, each locus’s former, center, and latter domains are initialized according to the discrete label;
Once all loci are initialized, they are assembled into genes;
The JudgeGene function then assesses the gene’s rationale. If deemed suitable, the next chromosome’s initialization is performed, and so forth until the entire chromosome population is processed. Should the gene fail the check, the chromosome is re-initialized. Taking chromosome I as an example, it is properly initialized because the center area at all loci is not all 0 and the latter area is all 0 or all 1.
Chromosome selection, crossover, mutation
In this study, we introduce the DGAARM, which incorporates a unique roulette wheel-based strategy in Genetic Algorithm (GA). Unlike traditional approaches, our strategy favors chromosomes with smaller fitness values, as determined by Eq (4).
| (4) |
In this context, w1, w2, and w3 denote weights such that their sum equals one. Chromosomes with smaller fitness values possess a higher probability of undergoing subsequent crossover and mutation operations.
The selection algorithm proceeds as follows and is illustrated in Fig 5:
Fig 5. Process of chromosome selection.
Compute the fitness value for each chromosome within the population. Using the inverse of these values, construct a simulated roulette wheel.
ii)Generate a random selection probability. If this probability falls within the interval of a particular chromosome on the wheel, that chromosome is chosen.
Assume that the fitness of chromosomes I, II, and III are 0.3,0.6, and 0.1, respectively. Then, calculate the ratio of their reciprocal to the sum of all and they are about 22%,10%, and 66%. We add them in turn to get the value on the roulette wheel and finally, according to the random number 0.48 generated by the roulette pointer, we select chromosome III.
The crossover operation is invoked when a randomly generated number, pCross, exceeds the current crossover rate, which is dynamic and varies across generations. At the onset of the algorithm, a high crossover rate is essential to rapidly identify feasible solutions within the solution space. As the algorithm progresses, this rate is marginally reduced to refine feasible solutions and preserve superior chromosome segments. This dynamic crossover rate is encapsulated by Eq (5).
| (5) |
Where currentIterateNum, totalIterateNum, and crossRate represent the current iteration round, total iteration rounds, and initial crossover rate, respectively.
For the crossover operation, two chromosomes are selected via the aforementioned selection algorithm. Subsequently, information on congruent loci in the genes of these chromosomes is exchanged.Crossover Process:
If pCross exceeds the current crossover rate, proceed; else, initiate mutation.
Generate crossover points randomly.
Via the selection algorithm, two chromosomes are chosen. The prior, central, and posterior domains at the identified crossover sites undergo an exchange of information.
Assess the viability of the post-crossover chromosome genes. If they’re viable, finalize the crossover and update the population. Otherwise, revert to the pre-crossover state.
Mutation process ensues post-crossover, triggered exclusively when a randomly generated number pChange surpasses the present mutation rate. This rate is dynamic, initiated at the start of the algorithm and then reducing non-linearly over time. This fluctuation is characterized by Eq (6).
| (6) |
Where currentIterateNum, totalIterateNum, and changeRate represent the current iteration, total iterations, and initial mutation rate, respectively.
In our study, we employ single-point mutation, implying that a single locus is arbitrarily chosen for informational modification to enact the mutation. The procedure is:
If pChange is above the ongoing variation rate, proceed; else, halt the mutation.
Arbitrarily pinpoint mutation sites.
Utilizing the selection algorithm, mutations are induced in the prior, central, and posterior domains of the chosen chromosomal mutation site. Specifically, the prior domain undergoes random feature selection from the current discrete feature set, storing its index. Meanwhile, the central and posterior domains select values from the {0,1} set.
The viability of the mutated chromosome genes is assessed. If deemed viable, the mutation is finalized and the population data updated. If not, revert to the pre-mutation state.
Renewal iterations of chromosome populations
The iterative update is a cornerstone of the algorithm’s control logic. Its purpose is to ensure the algorithm consistently converges towards an optimal chromosome. To facilitate this, the most favorable chromosome from the preceding generation is retained within the current generation’s population. This strategic retention either promotes convergence towards that particular chromosome or incentivizes the search for a superior one. If the entire chromosome population of the current generation hasn’t undergone an update, the selection algorithm is invoked. This identifies a chromosome from the preceding generation’s population, subjecting it to crossover and mutation operations when corresponding rates are met. This continues until the entire current chromosome population has been iteratively updated. The process of renewal iteration is delineated below:
Increment the iteration count.
- For each chromosome in the prior generation’s population:
- Execute selection, crossover, and mutation operations.
- Assess if the fitness value of the newly formed chromosome surpasses the minimum fitness value of the preceding generation’s chromosomes. If superior, retain the new chromosome. Otherwise, maintain the original.
Evaluate if the fitness value of the new chromosome exceeds the minimum fitness value from the prior generation. If so, keep the new chromosome; if not, revert to the original.
Construct a new generation of chromosome populations.
Results and discussion
Experiments detailed herein were conducted on a system featuring Windows 10, an Intel® Core™ i5-7200U CPU (2.50GHz to 2.70GHz), and 8GB of RAM, with implementations in Python. In order to verify the effectiveness of the well-conceived gene coding in spatial storage, two datasets were employed additionally: the nursery dataset and breast-cancer dataset sourced from the University of California’s UCI Machine Learning Repository. Pertinent characteristics of both datasets can be found in Tables 4 and 5.
Table 4. Nursery dataset related notes.
| Description items | Parameter Value |
|---|---|
| Number of data bars | 12960 |
| Number of attributes | 9 |
| Attribute 1—parents | usual, pretentious, great_pret |
| Attribute 2 -nursery | proper, less_proper, improper, critical, very_crit |
| Attribute 3 -family | complete, completed, incomplete, foster |
| Attribute 4—child | 1, 2, 3, more |
| Attribute 5—housing | convenient, less_conv, critical |
| Attribute 6—finance | convenient, inconv |
| Attribute 7—society | non-prob, slightly_prob, problematic |
| Attribute 8—health | recommended, priority, not_recom |
| Attribute 9—recommended results | not_recom, recommend, very_recom, priority, spec_prior |
Table 5. Breast-cancer dataset related notes.
| Description items | Parameter Value |
|---|---|
| Number of data bars | 286 |
| Number of attributes | 10 |
| Attribute 1—class | no-recurrence-events, recurrence-events |
| Attribute 2 -age | 10–19, 20–29, 30–39, 40–49, 50–59, 60–69,70–79,80–89,90–99 |
| Attribute 3 -menopause | >40, <40, premeno |
| Attribute 4—tumor-size | 0–4, 5–9, 10–14, 15–19, 20–24, 25–29, 30–34, 35–39, 40–44, 45–49, 50–54, 55–59 |
| Attribute 5—inv-nodes | 0–2, 3–5, 6–8, 9–11, 12–14, 15–17, 18–20, 21–23, 24–26, 27–29, 30–32, 33–35, 36–39 |
| Attribute 6—node-caps | yes, no |
| Attribute 7—deg-malig | 1, 2, 3 |
| Attribute 8—breast | left, right |
| Attribute 9—breast-quad | left-up, left-low, right-up, right-low, central |
| Attribute 10—irradiat | Yes, no |
Nursery database was derived from a hierarchical decision model originally developed to rank applications for nursery schools. It was used during several years in 1980’s when there was excessive enrollment to these schools in Ljubljana, Slovenia, and the rejected applications frequently needed an objective explanation. This dataset has a total of 12,960 records, each of which contains a total of 9 attributes, each with a different attribute value. The first eight attribute values have some correlation with the last attribute value.
Breast-cancer dataset was obtained from the University Medical Centre, Institute of Oncology, Ljubljana, Yugoslavia. This dataset has a total of 286 records, each of which contains a total of 10 attributes, each with a different attribute value. The first nine attribute values have some correlation with the last attribute value.
Our proposed DGAARM exhibits exemplary optimization concerning chromosome gene coding space storage, rule mining quantity, algorithm convergence, and rule quality. These findings offer invaluable insights for scholars exploring air quality determinants and furnish robust technological support for mining association rules in diverse application domains. Comprehensive comparative results and discussions are elucidated in subsequent sections.
Comparison of chromosomal gene coding space storage
A salient distinction in chromosomal genetic code space storage stems from the categorization status of dataset items. Grouping items permits the consolidation of non-sequitur related or similar-category items within a shared storage, negating the necessity for discrete storage spaces. Such a coding paradigm effectively diminishes the redundancy between analogous items, optimizing space usage, which delineated in Fig 6. Since the specific categories under the features of the Beijing dataset are more than those of the other two datasets, the space can be reduced more significantly.
Fig 6. Comparison of chromosomal gene coding space storage before and after data set category classification.
Rule number mining comparison
In the aspect of rule number mining, we find that because traditional association rule algorithms are affected by manual support settings, the number of rules mined under different support thresholds is different. When the support level is low, the number of rules that can be mined is large. As the support level increases, the number of rules that can be mined is significantly reduced, because the manual support settings filter out a large number of item sets. As we can see, when applying Apriori to Nursery data sets, a large number of rules can be mined when minimum support is set to 0.1 and minimum confidence is set to 0.2, while rules are no longer mined when minimum support is set to 0.2 and minimum confidence is set to 0.6. However, when DGAARM was applied to 10 repeated experiments on the Nursery dataset, it was found that by eliminating the step of setting the threshold parameter, it consistently looked for high-quality rules out of about 298 rules, as shown in Fig 7. Table 6 lists the rules on some Nursery datasets that are mined in part by DGAARM. For example, the family in society is problematic, then his financial aspect is inconvenient or families recommended for admission are financially convenient. We can derive the rules based on the weight of the evaluation, rather than filtering some item set by support thresholds
Fig 7. Comparison of the number of Apriori and DGAARM rule mining on the nursery dataset.
Table 6. Extracting association rules on the Nursery dataset using DGAARM with different weights of support, confidence, and lift.
| Rules | Support weight | Confidence weight | Lift weight | Interest level |
|---|---|---|---|---|
| {society-problem} = >{finance-inconvenient} | 0.1 | 0.9 | 0 | 0.46 |
| {result-recommended} = >{finance-convenient} | 0.1 | 0.9 | 0 | 0.90 |
| {family-foster care, health-healthy} = >{result-preferred} | 0.2 | 0.8 | 0 | 0.43 |
| {finance-convenient} = >{family-complete} | 0.3 | 0.4 | 0.3 | 0.28 |
| {finance-convenient} = >{result-preferred,child-3,family-incomplete,nursery-very critical} | 0.4 | 0.5 | 0.1 | 0.50 |
| {child-3, parents-ordinary} = > {finance-inconvenient} | 0.4 | 0.5 | 0.1 | 0.32 |
| {society-minor problem} = >{housing-inconvenient} | 0.5 | 0.4 | 0.1 | 0.22 |
Similarly, we can see from Fig 8 that when Apriori is applied to the Beijing dataset, a large number of rules can be mined when the minimum support is set to 0.5 and the minimum confidence is set to 0.5, while the minimum rules are mined when the minimum support is set to 0.7 and the minimum confidence is set to 0.9. However, when DGAARM was applied to 10 repeated experiments on the Beijing dataset, it could consistently look for high-quality rules from about 318 rules.
Fig 8. Comparison of the number of Apriori and DGAARM rule mining on the Beijing dataset.
Comparison between DGAARM and other rule mining methods
This experiment provides a comprehensive comparison of the DGAARM algorithm with a range of traditional association rule mining algorithms including Apriori, FP-Growth, and Eclat. In addition, we also compare rule mining algorithms that integrate swarm intelligence, such as Particle Swarm Association Rule Mining (PSOARM), Multi-objective Particle Swarm Association Rule Mining (MOPSOARM), Whale Association Rule Mining (WOAARM), Differential Evolution Association Rule Mining (DEARM), as well as Animal Dynamic Migration Association Rule Mining (ADMOARM).
Key performance indicators employed in this comparison include rule mining time consumption, with pertinent results presented in Table 7. During the testing phase, algorithmic parameters such as support, confidence, and lift were calibrated at weights of 0.3, 0.6, and 0.1 respectively. Each algorithm was executed ten times, with individual run times recorded and visually represented in Fig 9.
Table 7. Comparison of running time between DGAARM algorithm and Apriori and other 8 algorithms on the Beijing data set.
| Methods | Time (s) |
|---|---|
| Apriori | 156.322 |
| FP-Growth | 148.804 |
| Eclat | 211.027 |
| PSOARM | 313.510 |
| MOPSOARM | 320.008 |
| WOAARM | 352.201 |
| DEARM | 302.129 |
| ADMOARM | 270.301 |
| DGAARM (ours) | 249.813 |
Fig 9. Single and average running time of the DGAARM algorithm for 10 times.
Insights derived from Table 7 indicate that DGAARM’s runtime marginally surpasses that of conventional rule mining algorithms. This variance can be attributed to the latter’s approach of setting minimum thresholds and consequently filtering specific items, optimizing run times. Remarkably, when juxtaposed with swarm intelligence-optimized rule mining algorithms, DGAARM boasts the most efficient execution time. Such efficiency stems from its superior coding methodology combined with a dynamic search strategy, specifically its dynamic crossover and mutation rates, detailed further in Fig 10.
Fig 10. Comparison of raw and dynamic crossover/mutation rate change process.
To illustrate the advantages of DGAARM in terms of convergence performance, we set the number of iterations of each algorithm to 30 and record the interesting-ness value of the optimal rule in each iteration. Inspection of Fig 11 reveals that the algorithm’s performance convergence aligns well with theoretical expectations. Specifically, during the initial phases, DGAARM rapidly identifies rules of high interest when both the crossover and mutation rates are elevated. As these rates stabilize at lower values in subsequent stages, the algorithm converges the interest values of the rules to a stable equilibrium. Notably, due to its dynamic search strategy, DGAARM achieves a more efficient convergence time compared to other algorithms such as PSOARM, MOPSOARM, WOAARM, and DEARM.
Fig 11. Comparison of convergence of the DGAARM algorithm with the other five optimization algorithms after 30 iterations.
Rule quality analysis
DGAARM is engineered to discern rules of interest that yield superior quality. In this paper, a unique approach is put forward wherein varying weight values for support, confidence, and lift serve as a fitness function, thus enabling the extraction of rules with superior function values. Table 8 lists some of the rules mined using DGAARM. When focusing on rule confidence, for instance, we can get the rule {temperature_warm} = > {barometric pressure_low}. This is coherent with the environmental reality that balmy temperatures correlate with lower barometric pressures.
Table 8. Extracting air quality influencing factor association rules using DGAARM in Beijing meteorological dataset.
| Rules | Support weight | Confidence weight | Lift weight | Interest level |
|---|---|---|---|---|
| {O3_II, Temperature_Warm} = >{SO2_I, Rainfall_R0} | 0.1 | 0.9 | 0 | 0.9 |
| {Temperature_Warm} = >{Barometric pressure_Low pressure} | 0.1 | 0.9 | 0 | 0.79 |
| {AQI-5, Wind power_Force 3 wind, Wind direction_Northeast} = >{Barometric pressure_High pressure, Relative humidity_Normal, Rainfall_R0} | 0.1 | 0.9 | 0 | 0.20 |
| {Relative humidity_Humidity, Wind direction_Northeast} = >{Rainfall_R1, Wind power_Force 3 wind} | 0.2 | 0.2 | 0.6 | 0.60 |
| {AQI-1} = >{PM10_I, Temperature_Warm, Relative humidity_Humidity} | 0.2 | 0.2 | 0.6 | 0.60 |
| {AQI-5} = >{CO_II, Weather condition_Overcast, Temperature_Lightly cold, Wind power_Force 3 wind} | 0.2 | 0.2 | 0.6 | 0.60 |
| {Barometric pressure_High pressure} = >{Rainfall_R0} | 0.3 | 0.3 | 0.4 | 0.62 |
| {AQI-1, Barometric pressure_High pressure} = >{O3_I} | 0.3 | 0.3 | 0.4 | 0.44 |
| {NO2_III, PM10_II, Barometric pressure_High pressure, Temperature_Slightly cold} = >{SO2_I} | 0.4 | 0.4 | 0.2 | 0.36 |
| {Barometric pressure_High pressure} = >{PM10_I,PM2_5_I} | 0.4 | 0.4 | 0.2 | 0.30 |
| {O3_I} = >{SO2_I} | 0.5 | 0.5 | 0 | 0.87 |
| {CO_I, Relative humidity_Normal, Wind power_Light breeze} = >{NO2_II} | 0.5 | 0.5 | 0 | 0.32 |
| {Weather condition_Sunny, Barometric pressure_Low pressure, Temperature_Warm, Wind power_Force 3 wind} = >{SO2_I} | 0.7 | 0.2 | 0.1 | 0.20 |
| {CO_I, SO2_I, Barometric pressure_Low pressure} = >{O3_II} | 0.7 | 0.2 | 0.1 | 0.20 |
| {AQI-1, PM2_5_I} = >{Barometric pressure_High pressure} | 0.8 | 0.1 | 0.1 | 0.25 |
| {NO2_V, Relative humidity_Humidity} = >{CO_III, SO2_III, Barometric pressure_High pressure} | 0.8 | 0.1 | 0.1 | 0.20 |
Similarly, when emphasizing rule lift, the algorithm unveils the rule {AQI-1} = > {PM10_I, temperature_mild, relative_humidity}. This illustrates the scenario when excellent air quality is accompanied by minimal PM10 concentrations, mild temperatures, and relatively humid conditions.
In instances where support is the emphasis, akin to classical algorithms such as Apriori, it is possible to derive the rule {O3_I} = >{SO2_I}. This suggests that both ozone and SO2 concentrations are classified at rank one. In addition, Table 9 lists six items with low support in the Beijing data set. When the traditional method such as Apriori is used to set the support level to 0.2, it is impossible to mine the rules, but these rules can be found in Table 8, such as {O3_II, Temperature_Warm} = > {SO2_I, Rainfall_R0}, {AQI_V} = > {CO_II, Weather condition_Overcast, Temperature_Lightly cold, Wind power_Force3 wind}, etc.
Table 9. Six low support items in Beijing meteorological dataset.
| Item | support |
|---|---|
| Temperature_Lightly cold | 0.1113 |
| O3_II | 0.1126 |
| Temperature_Slightly cold | 0.1230 |
| Wind power_Force 3 wind | 0.1539 |
| NO2_III | 0.1686 |
| Relative humidity_Humidity | 0.1766 |
It is critical, however, to note that excessive concentration on the aspect of support could lead to the extraction of rules with markedly reduced interest levels, potentially yielding results of low relevance to the researcher.
This resonates with the issue that conventional association rules might fail to identify pertinent rules when the support level is high. In such scenarios, redirecting focus onto other evaluative metrics via the DGAARM algorithm could prove highly beneficial.
Statistical analysis
In order to better show that our algorithm is not limited by the minimum support threshold and can mine the rules with low support, a statistical analysis is carried out. The steps are as follows: First we establish the null hypothesis, which is the argument described above, and by rejecting this null hypothesis we can statistically prove our validity. Secondly, we conducted random statistics on the support degree of rules mined by DGAARM and traditional algorithm Apriori on the Beijing dataset, and then conducted Student’s t test, which is a statistical method used to measure the deviation degree of observed values from expected values. Table 10 shows the test results. The value for μ, standard deviation (s), t-value (t) and p-value of each algorithm are obtained from the t-test, where μ represents the average support degree of the mined rule, and s represents its standard deviation of the rules mined. The standard deviation is calculated by Eq (7), the t-value is calculated by Eq (8), and n1 and n2 are the number of random sample rules, which are set as 10.
Table 10. Student’s t test of DGAARM and Apriori.
| Beijing(support = 0.3) | ||||
|---|---|---|---|---|
| μ | s | t | p-Value | |
| DGAARM | 0.265 | 0.0193 | 3.326 | 0.00375 |
| Apriori | 0.561 | 0.0522 | ||
| (7) |
| (8) |
The p-value can be calculated by subtracting the normal distribution value of t from 1. If the p-value is lower than a given significance level α, the established null hypothesis can be rejected. In this experiment, we set the significance level α as 0.05 (5%). In Table 10, a significance level of p-value less than 0.05 can be observed. Therefore, we can determine that there is a statistically significant difference between DGAARM and the traditional rule Apriori in mining rules that are exempt from support threshold setting. In other words, we can mine rules that are exempt from support thresholds.
Conclusions
With the gradual deepening of the process of economic globalization, environmental problems brought by rapid economic development have attracted more and more attention. Good air quality is crucial to people’s physical and mental health and social activities. In addition, the current global climate change is accelerating, and extreme weather conditions are also having a huge impact on air quality. Under the double influence, it is of great significance to explore the correlation between air quality and meteorological sources. This manuscript innovatively proposes DGAARM based on traditional genetic algorithm and association rule mining technology, and applies it to air pollution correlation analysis, which can effectively reveal the correlation between air quality and various factors at different levels. Key findings from the experimental outcomes highlight:
The design of a novel gene coding strategy that is rooted in a single locus and capable of carrying multiple information chromosomes reduced the rule space storage consumption by 50%.
The dynamic crossover and mutation rate are proposed in the process of optimal search, which makes the algorithm have strong global search ability in the initial execution, and transition to fast convergence in the subsequent algorithm iteration. Based on it and the special coding strategy above, we reduced about 20% in terms of search time compared with some heuristic algorithm, while improving convergence.
The implementation of the algorithm’s new evaluation index is not limited by the threshold of support and confidence, and can stably mine the association rules, whose interest level can be up to 90% between frequent and infrequent items in the object database.
DGAARM can complete air quality impact factor mining after preprocessing complex Beijing data sets including discrete and continuous data. Future research will consider adapting more different types of air quality data in the future, as well as integrating clustering technology into the data preprocessing stage to increase mining at different feature levels.
Acknowledgments
The authors are grateful to the anonymous reviewers for their time and insightful comments.
Data Availability
The data supporting the research of the paper is stored in the stable repository github, which can be accessed through https://github.com/academicWq/essay/.
Funding Statement
We receive funding from The Universities Outstanding Outstanding Talents Project of Anhui Provincial (grant number gxyq2022075) and The Open Key Laboratory Project of Anhui Provincial (grant number IBES2021KF07),And the author who received the two award are Xiaoxuan Wu. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Agrawal R, Imielinski T, Swami A. Mining association rules between sets of items in large databases[C]//Acm sigmod record.ACM, 1993,22(2):207–216. [Google Scholar]
- 2.José A. Delgado-Osuna, Carlos García-Martínez, José Gómez-Barbadillo, Sebastián Ventura, Heuristics for interesting class association rule mining a colorectal cancer database, Information Processing &Management, Volume 57, Issue 3,2020,102207, ISSN 0306-4573, 10.1016/j.ipm.2020.102207. [DOI] [Google Scholar]
- 3.Elif Varol Altay, Bilal Alatas, Association analysis of Parkinson disease with vocal change characteristics using multi-objective metaheuristic optimization, Medical Hypotheses, Volume 141, 2020, 109722, ISSN 0306-9877, doi: 10.1016/j.mehy.2020.109722 [DOI] [PubMed] [Google Scholar]
- 4.Erkan Çakır, Remzi Fışkın, Coşkan Sevg ili, Investigation of tugboat accidents severity: An application of association rule mining algorithms, Reliability Engineering & System Safety, Volume 209,2021,107470, ISSN 0951-8320, 10.1016/j.ress.2021.107470. [DOI] [Google Scholar]
- 5.Verma Abhishek, Dhalmahapatra Krantiraditya, J. Maiti, Forecasting occupational safety performance and mining text-based association rules for incident occurrences, Safety Science, Volume 159,2023,106014, ISSN 0925-7535, 10.1016/j.ssci.2022.106014. [DOI] [Google Scholar]
- 6.Jungyeol Hong, Tamakloe Reuben, Park Dongjoo, Application of association rules mining algorithm for hazardous materials transportation crashes on expressway, Accident Analysis & Prevention, Volume 142, 2020, 105497, ISSN 0001-4575, 10.1016/j.aap.2020.105497. [DOI] [PubMed] [Google Scholar]
- 7.Liu Jiangyan, Shi Daliang, Li Guannan, Xie Yi, Li Kuining, Liu Bin, et al, Data-driven and association rule mining-based fault diagnosis and action mechanism analysis for building chillers, Energy and Buildings, Volume 216, 2020, 109957, ISSN 0378-7788, 10.1016/j.enbuild.2020.109957. [DOI] [Google Scholar]
- 8.Agrawal R., Imieliński T., & Swami A. (1993). Mining association rules between sets of items in large databases. In Proceedings of the 1993 ACM SIGMOD international conference on management of data (pp. 207–216). 10.1145/170035.170072. [DOI] [Google Scholar]
- 9.Huang Daozheng, Liang Tiantian, Hu Shenping, Loughney Sean, Wang Jin, Characteristics analysis of intercontinental sea accidents using weighted association rule mining: Evidence from the Mediterranean Sea and Black Sea, Ocean Engineering, Volume 287, Part 2, 2023, 115839, ISSN 0029-8018, 10.1016/j.oceaneng.2023.115839. [DOI] [Google Scholar]
- 10.Agrawal R., Srikant R., 1994. Fast algorithms for mining association rules. In: Proc.20th International Conference on Very Large Data Bases, VLDB, vol. 1215, pp. 487–499. [Google Scholar]
- 11.Han J, Pei J, Yin Y. Mining frequent patterns without candidate generation[C]//ACM sigmod record.ACM,2000,29(2):1–12. [Google Scholar]
- 12.Zhang C, Tian P, Zhang X, et al. Fast Eclat Algorithms Based on Minwise Hashing for Large Scale Transactions. IEEE Internet of Things Journal. 2019; PP(99):3948–3961. [Google Scholar]
- 13.Wang Lei, Guo Yan, Guo Yuyao, Xia Xuhui, Zhang Zelin, Cao Jianhua, An Improved Eclat Algorithm based Association Rules Mining Method for Failure Status Information and Remanufacturing Machining Schemes of Retired Products, Procedia CIRP, Volume 118, 2023, Pages 572–577, ISSN 2212-8271, 10.1016/j.procir.2023.06.098. [DOI] [Google Scholar]
- 14.Liu D H, Zhao X, Zhao J P, et al. Dynamic association network analysis of accident causes of tower crane based on improved Apriori algorithm[J]. Safety and Environmental Engineering, 2023,30(6):106–112. (in Chinese) [Google Scholar]
- 15.YU Biao LIU Shou- quan. Improved Association Rules Mining Algorithm Based on FP-growth Algorithm[J]. Computer&Network, 2017, 43(14): 68–71. (in Chinese) [Google Scholar]
- 16.Wang Huan-Bin, Gao Yang-Jun, Research on parallelization of Apriori algorithm in association rule mining, Procedia Computer Science, Volume 183,2021, Pages 6641–647, ISSN 1877-0509, 10.1016/j.procs. 2021.02.109. [DOI] [Google Scholar]
- 17.Liu L, Wen J, Zheng Z, Su H. An improved approach for mining association rules in parallel using Spark Streaming. Int J Circ Theor Appl. 2021; 49:1028–1039. 10.1002/cta.2935. [DOI] [Google Scholar]
- 18.Goldberg D.E., 1989. Genetic Algorithms in Search, Optimization and Machine Learning, first ed. Addison-Wesley Longman Publishing Co. Inc., USA. [Google Scholar]
- 19.Zhu Jianlu, Wang Guocong, Li Yuxing, Duo Zhili, Sun Chongzheng, Optimization of hydrogen liquefaction process based on parallel genetic algorithm, International Journal of Hydrogen Energy, Volume 47, Issue 63, 2022, Pages 27038–27048, ISSN 0360-3199, 10.1016/j.ijhydene.2022.06.062. [DOI] [Google Scholar]
- 20.Cui Huixia, Qiu Jianlong, Cao Jinde, Guo Ming, Chen Xiangyong, Gorbachev Sergey, Route optimization in township logistics distribution considering customer satisfaction based on adaptive genetic algorithm, Mathematics and Computers in Simulation, Volume 204, 2023, Pages 28–42, ISSN 0378-4754, 10.1016/j.matcom.2022.05.020. [DOI] [Google Scholar]
- 21.Sun Kexin, Zheng Debin, Song Haohao, Cheng Zhiwen, Lang Xudong, Yuan Weidong, et al. , Hybrid genetic algorithm with variable neighborhood search for flexible job shop scheduling problem in a machining system, Expert Systems with Applications, Volume 215, 2023, 119359, ISSN 0957-4174, 10.1016/j.eswa.2022.119359. [DOI] [Google Scholar]
- 22.Sharmila S., & Vijayarani S. (2021). Association rule mining using fuzzy logic and whale optimization algorithm. Soft Computing, 25(2), 1431–1446. 10.1007/s00500-020-05229-4. [DOI] [Google Scholar]
- 23.Elif Varol Altay, Bilal Alatas, Differential evolution and sine cosine algorithm based novel hybrid multi-objective approaches for numerical association rule mining, Information Sciences, Volume 554, 2021, Pages 198–221, ISSN 0020-0255, 10.1016/j.ins.2020.12.055. [DOI] [Google Scholar]
- 24.Subramanian Kannimuthu and Kandhasamy Premalatha. ‘Mining High Utility Itemsets Using Genetic Algorithm Based-Particle Swarm Optimization (GA-PSO)’. 1 Jan. 2023: 1169–1189. [Google Scholar]
- 25.Kannimuthu S., Chakravarthy D.G. Discovery of Interesting Itemsets for Web Service Composition Using Hybrid Genetic Algorithm. Neural Process Lett 54, 3913–3939 (2022). 10.1007/s11063-022-10793-x. [DOI] [Google Scholar]
- 26.Tyagi S., & Bharadwaj K. K. (2013). Enhancing collaborative filtering recommendations by utilizing multi-objective particle swarm optimization embedded association rule mining. Swarm and Evolutionary Computation, 13, 1–12. 10.1016/j.swevo.2013.07.001. [DOI] [Google Scholar]
- 27.Hu Kerui, Qiu Lemiao, Zhang Shuyou, Wang Zili, Fang Naiyu, An animal dynamic migration optimization method for directional association rule mining, Expert Systems with Applications,Volume 211, 2023, 118617, ISSN 0957-4174, 10.1016/j.eswa.2022.118617. [DOI] [Google Scholar]
- 28.Zhijin Li, Yanbin Hao. Meta-analysis of the Relationship Between PM2.5, PM10, O3 Concentrations and Meteorological Conditions in Multiple Cities [J]. Journal of Green Science and Technology. 2022, 24(02):110–113+125. doi: 10.16663/j.cnki.lskj 2022.02.005.(in Chinese) [DOI] [Google Scholar]
- 29.Xiaoyue Zhu. Study on Air Pollution Characteristics and Influencing Factors in Jiangsu Province Based on Grey Relational Analysis [D]. Nanjing University of Aeronautics and Astronautics, 2018. (in Chinese) [Google Scholar]
- 30.Duan Wenjiao, Wang Xiaoqi, Cheng Shuiyuan, Wang Ruipeng, A new scheme of PM2.5 and O3 control strategies with the integration of SOM, GA and WRF-CAMx, Journal of Environmental Sciences, Volume 138, 2024, Pages 249–265, ISSN 1001-0742, 10.1016/j.jes.2023.02.058 [DOI] [PubMed] [Google Scholar]
- 31.Geng Xiaojiao, Liang Yan, Jiao Lianmeng, ARC-SL: Association rule-based classification with soft labels, Knowledge-Based Systems, Volume 225, 2021, 107116, ISSN 0950-7051, 10.1016/j.knosys.2021.107116. [DOI] [Google Scholar]
- 32.Zeynab Fallah Sokhangoee Abdoreza Rezapour, A novel approach for spam detection based on association rule mining and genetic algorithm, Computers & Electrical Engineering, Volume 97,2022, 107655, ISSN 0045-7906, 10.1016/j.compeleceng.2021.107655. [DOI] [Google Scholar]
- 33.Ruihua Xu, Fan Luo, Risk prediction and early warning for air traffic controllers’ unsafe acts using association rule mining and random forest, Safety Science, Volume 135, 2021, 105125, ISSN 0925-7535, 10.1016/j.ssci.2020.105125. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data supporting the research of the paper is stored in the stable repository github, which can be accessed through https://github.com/academicWq/essay/.











