Connect with us

AI Insights

A deep dive into artificial intelligence with enhanced optimization-based security breach detection in internet of health things enabled smart city environment

Published

on


This paper proposes a novel SADDBN-AMOA technique for smart city-based IoHT networks. The main aim of the SADDBN-AMOA technique is to provide a resilient attack detection method in the IoHT environment of smart cities to mitigate security threats. It contains data pre-processing, feature optimizer, DL with attack detection, and model tuning with IHHO. Figure 2 depicts the entire process of the SADDBN-AMOA model.

Fig. 2

Overall process of SADDBN-AMOA model.

Data Pre-processing through normalization

The data pre-processing step initially applies the Z-score normalization method for transforming input data into a structured pattern31. This method is chosen for its efficiency in standardizing data by converting features to have a mean of zero and a standard deviation of one, which assists in stabilizing the training process and speeds up convergence. Unlike min-max scaling, which can be sensitive to outliers, Z-score normalization is more robust as it accounts for data distribution, mitigating the impact of extreme values. This method ensures that features with diverse units or scales contribute equally to the model, averting bias towards variables with larger ranges. Additionally, it enhances the performance of models such as DBN, which usually assume distributed input data. Hence, this improves the model’s stability, accuracy, and generalization compared to other scaling techniques.

Z-score normalization, or standard score normalization, is a data pre-processing model that converts feature values by centring them near the mean with a standard deviation of 1. Concerning attack detection in IoHT atmospheres, this model assistances guarantees that each feature corresponds to the detection method, particularly after they have dissimilar units or scales. This is important for ML methods aware of feature sizes, like neural networks, SVM, or k-NN. This model improves model accuracy, performance, and convergence speed by decreasing bias from leading features. In IoHT methods, while sensor data might differ extensively, this standardization enhances the dependability of intrusion detection or anomaly. It also facilitates more efficient classification and feature selection in resource-restricted IoHT devices.

Feature optimization with SMO model

For selecting the feature process, the proposed SADDBN-AMOA model utilizes the SMO method to choose the most related features from the data32. This model was selected for its robust global search capability and efficient exploration-exploitation balance, which assists in avoiding premature convergence, which is common in other algorithms. Unlike conventional methods like genetic algorithms (GA) or particle swarm optimization (PSO), SMO illustrates faster convergence and improved adaptability to intrinsic, high-dimensional data. Its bio-inspired mechanisms effectively detect the most relevant features, mitigating dimensionality without losing critical data. This results in an enhanced accuracy of the model and reduced computational cost. Moreover, the simplicity and fewer parameters of the model make it easier to implement and tune, presenting a robust and reliable approach for optimizing feature subsets in IoHT security applications. SMO is chosen due to its latest success as a meta-heuristic model stimulated by foraging behaviour and slime mould dispersion. The SMO has three stages: wrap, approach, and search for food. The mathematical representation of all expressions is mentioned below.

Stage 1-Approaching Food: Subject to the smell (scent) that food creates, SM might approach it. Like the technique of using mathematical equations to observe the mode of contraction approaching behaviours, the succeeding equations are presented to mimic it. The equation that inspires the SM to approach food is calculated by Eq. (1)

$$:overrightarrow{X}left(t+1right)=left{begin{array}{ll}{overrightarrow{X}}_{b}left(tright)+overrightarrow{{v}_{b}}times:left(overrightarrow{W}times:{overrightarrow{X}}_{A}left(tright)-{overrightarrow{X}}_{B}left(tright)right),&:r

(1)

Here, (:overrightarrow{{v}_{b}}) refers to parameters ranging from (:-1) to (:+1), and (:overrightarrow{{v}_{c}}) linearly reduces from 1 to (:0) for the last specific iteration. (:W) denotes the slime mould’s weight, and (:t) represents the iteration that is now in progress. From the SM, the (:{overrightarrow{X}}_{A}left(tright)) and (:{overrightarrow{X}}_{B}left(tright)) are selected randomly. The (:overrightarrow{X}) characterizes the location of the SM, while (:{overrightarrow{X}}_{b}) specifies the present individual’s location at the odour focus of the meal is greater. Equation (2) is utilized for calculating the (:p)-value.

$$:p=text{t}text{a}text{n}text{h}left|Sleft(iright)-DFright|:$$

(2)

The fitness of (:overrightarrow{X}) was characterized by (:Sleft(iright)), whereas the maximal fitness gained through each iteration (:S(iin:text{1,2},3dots:dots:dots:n)) is signified by DF. The (:overrightarrow{{v}_{b}}) value varies from − a to (:a), and the (:a) value is calculated using Eqs. (3) and (4).

$$:overrightarrow{{v}_{b}}in:left[-a:to:aright]$$

(3)

$$:a=arctext{t}text{a}text{n}text{h}left(-left(frac{t}{{text{m}text{a}text{x}}_{iter}}right)+1right)$$

(4)

The weight ((:W)) value is measured by Eq. (5) as shown,

$$vec{W}~left( {SmellIndexleft( i right)} right) = left{ {begin{array}{*{20}l} {1 + r cdot logleft( {frac{{b_{F} – Sleft( i right)}}{{b_{F} – w_{F} }} + 1} right),} hfill & {if,Sleft( i right) < Medleft[ S right]} hfill \ {1 – r cdot logleft( {frac{{b_{F} – Sleft( i right)}}{{b_{F} – w_{F} }} + 1} right),} hfill & {otherwise} hfill \ end{array} } right.$$

(5)

$$:SmellIndexleft(SIright)=Sortleft(Sright)$$

(6)

The above equation has the following elements: (:r) embodies randomly generated values within the range (:left[text{0,1}right]), which specifies that (:Sleft(iright)) is acknowledged for ranking in the higher portion. The (:{b}_{f:})and (:{w}_{f}) characterize the best and worst values of fitness gained within the present iterative method. The (:{text{m}text{a}text{x}}_{iter}) specifies the maximal iteration amount. (:Medleft[Sright])refers to median operator representation. The (:SmellIndexleft(SIright))displays the values of fitness in sequential order.

Stage 2-Food Wrapping: The quality and nature of the SM search model are impacted. A position’s weight will rise after a considerable focus on food is greater. It is compelled to search another area when the attention is lower as the area’s weight falls. This stage consists of upgrading the position of the SM utilizing Eq. (7):

$$:overrightarrow{X}left(t+1right)=left{begin{array}{ll}randcdot:left(UB-LBright)+LB,&:rand

(7)

Whereas (:z) denotes probability applied to attack a balance between exploitation and exploration, and (:UB) and (:LB) represent upper and lower bound limits.

Stage 3-Food Grabbling: SM moves to places with high food attention. (:W,) (:overrightarrow{{v}_{b}},) and (:overrightarrow{{v}_{c}}) characterize changing venous widths. (:overrightarrow{{v}_{b}}) and (:overrightarrow{{v}_{b}}) swing correspondingly between ([-a, a]) and [-1, 1]. For example, (:overrightarrow{{v}_{b}}) and (:overrightarrow{{v}_{b}}) become nearer to 0 when iteration improves.

The fitness function (FF) deliberates the classification accuracy and the preferred features. It maximizes the classification accuracy and reduces the set size of the selected attributes. Then, the succeeding FF is deployed to approximate unique results, as presented in Eq. (8).

$$:Fitness=alpha:cdot::ErrorRate+left(1-alpha:right)cdot:frac{#SF}{#All_F}$$

(8)

Now, (:ErrorRate) symbolizes the classification (:ErrorRate) utilizing the chosen features. (:ErrorRate) is calculated as the incorrect percentage categorized to the sum of classifications produced, identified as the value among (0,1), (:#SF) denotes preferred feature counts, and (:#All_F) stands for the complete quantity of features in the novel data set. (:alpha:) is used for controlling the prominence.

DL for attack detection

The DBN method is followed for the attack classification process33. This method is chosen for its effectual capability in learning hierarchical feature representations from complex and high-dimensional data, which is common in IoHT environments. Unlike conventional classifiers, DBN efficiently extracts deep features without extensive manual pre-processing, thus enhancing detection accuracy. The model also integrates unsupervised pretraining with fine-tuning, improving generalization and mitigating overfitting. Compared to models such as standard neural networks or SVMs, DBNs are more efficient in capturing complex patterns and dependencies in security breach data. This results in robust and reliable classification performance, making DBNs appropriate for the dynamic and diverse nature of IoHT security threats. Figure 3 demonstrates the structure of DBN.

Fig. 3
figure 3

DBNs are a DL method designed to capture composite hierarchic data representation. They include numerous stochastic, latent variables layers and can learn complex designs from higher-dimensional data. DBNs are valuable in settings, whereas feature learning and reducing the dimensions are essential, like network intrusion detection. This part examines the numerical basics, architecture, and DBNs application, highlighting their part in recognizing networking assaults. A DBN comprises numerous Restricted Boltzmann Machine (RBMs) layers and the supervised method’s last layer, often softmax classifiers. The DBN structure consists of the following elements:

RBMs: All RBMs in the DBN are two-layered, stochastic NN comprising a visible layer (VL) and a hidden layer (HL). The VL characterizes the input data, whereas the HL takes high-level characteristics. The network is trained to maximize the data probability, and the links among the layers are aimless. The main modules are:

  • VL: Characterizes the detected information and has real-valued or binary elements.

  • HL: Signifies the feature learning from the input information and takes concealed designs.

  • Weights: The relations between the HL and VL are characterized by weighting that is learned in training.

The RBM’s energy function is provided by:

$$:Eleft(nu:,:hright)=-{sum:}_{i}{b}_{i}{nu:}_{i}-{sum:}_{j}{c}_{j}{h}_{j}-{sum:}_{i,j}{w}_{ij}{nu:}_{i}{h}_{j}:$$

(9)

Whereas (:nu:) and (:h) represent hidden and visible elements, (:{w}_{ij}) are weighted, and(::{b}_{i}) and (:{c}_{j}) are biased.

Stacking RBMs: They are produced by stacking numerous RBMs in the deeper structure. All RBMs are trained layer-by-layer, whereas the output only RBM acts as the input to the following.

Last Layer of Classification: Afterward, RBM’s pretraining, the last supervised layer, is added to carry out regression or classification assignments. The learning procedure in DBNs consists of dual primary stages: fine-tuning and pretraining.

Pretraining: This stage consists of training all RBMs autonomously utilizing unsupervised learning. The aim is to learn the weights, which maximizes the data probability. The model of contrastive divergence is usually applied to train RBMs:

$$:varDelta:{w}_{ij}=eta:left(right{{v}_{i}{h}_{j}{}}_{data}-left{{v}_{i}{h}_{j}{}}_{model}right):$$

(10)

Here, (leftlangle {nu _{i} h_{j} } rightrangle _{{data}}) refers to the predictable value of the output from the information’s hidden and visible components, and (leftlangle {nu _{i} h_{j} } rightrangle _{{model}}) denotes the predicted value from the model distribution.

Finetuning: Afterward, the complete DBN is finetuned. This includes upgrading the weighting of the last layer of classification and, possibly, the weighting of the RBMs. This stage utilizes gradient descent to reduce the function of loss:

$$:L=-{sum:}_{i}({y}_{i}:text{l}text{o}text{g}left({widehat{y}}_{i}right)+left(1-{y}_{i}right)text{l}text{o}text{g}left(1-{widehat{y}}_{i}right):$$

(11)

Meanwhile, (:{y}_{i}) denotes true labels, and (:{widehat{y}}_{i}) represents forecast likelihoods.

DBNs may be used for learning the hierarchic representation of networking traffic data, allowing the recognition of composite attacking designs. The capability to remove and learn from higher-level attributes makes DBNs suitable for identifying unknown or known intrusions. By using DBNs, networking intrusion detection methods can take advantage of enhanced dimensionality reduction and feature extraction, improving their capability to recognize subtle patterns and anomalies suggestive of mischievous action. DBNs’ deeper hierarchical architecture permits further complex classification and understanding of networking traffic, possibly resulting in high precision and strength in identifying network attacks. Generally, DBNs provide a robust architecture to learn hierarchic feature representation. Their capability for modelling composite designs and decreasing dimensions makes them a helpful device in the current strength to improve cybersecurity through progressive DL methods.

Model tuning with IHHO approach

Eventually, the IHHO technique adjusts the DBN model’s hyperparameter values, resulting in greater classification performance34. This model is chosen for its superior exploration and exploitation balance, which effectively avoids local optima and ensures global search efficiency. This model enables dynamic adaptation and faster convergence compared to conventional optimization methods, namely GA or PSO, replicating the cooperative hunting strategy of Harris hawks. Its improved ability to fine-tune hyperparameters results in enhanced detection accuracy and mitigated false positives in complex IoHT environments. Moreover, the flexibility and robustness of the model make it appropriate for optimizing DL models, ensuring reliable performance even with high-dimensional and noisy data.

The HHO model is one of the population-based optimizer models that emulate the cooperative behaviour and hunting approach of the predacious bird HHO in the foraging process. Amongst the running of the model, the HHO has the candidate solution, and the target approach is the optimum solution with iterations. The HHO model might be segmented into 3 phases: local exploitation, conversion among exploitation and exploration, and global exploration.

The intention of the HHO model generates an issue with the following simplicity:

  • The tactic of arbitrarily initializing the population is employed. The arbitrary creation of the population results in the arbitrary distribution of created HHO individuals in the searching area, which results in the technique’s uncertainty.

  • In the method’s iterative procedure, only the data of the finest location is employed, which affects the complete model, falling into a local optimal when the existing finest location has a local optimal.

  • During the search process, the HHO model chooses the optimum acquiring approach completely reliant on the parameters (:lambda:) and (:E.) These dual parameters will instantly impact the HHO’s ending outcome. The decrease approach of escaping energy (:E) should attempt to fit the modification of the target’s energy in the procedure of escaping the actual situation, and the parameter value (:lambda:) additionally guarantees the unpredictability. Otherwise, it will decrease the model’s capability to exploit and search.

Logistic-tent chaotic mapping initializing populations

This model is accepted to resolve the irregular distribution of population initialization. The examination has depicted that integrating several lower-dimension chaotic mapping techniques to establish a complex, chaotic technique might effectively resolve the issues.

$$:{X}_{n+1}=mu:{X}_{n}left(1-{X}_{n}right),:n=text{1,2}cdots:$$

(12)

Here, (:mu:) represents the control parameter, and (:X) depicts the system variable. Logistic chaotic mapping has similar issues once the system parameters range from zero to one.

$$:{X}_{n+1}=left{begin{array}{l}frac{{X}_{n}}{alpha:},{X}_{n}in:left[0,:alpha:right)\:frac{(1-{X}_{n})}{(1-alpha:)},{X}_{n}in:left[alpha:,:1right]end{array}right.:$$

(13)

Here, X specifies the system variable and (:alpha:in:left(text{0,1}right)). It concerns some controlling parameters and restricted intervals of chaos.

According to the standard of Logistic-tent chaotic maps, the logistic and tent combined chaotic maps method might be formed by taking the outputs of Logistic maps as the input of Tent maps and iteratively analyzing them. This complex chaotic maps method shows the speed of iteration, and auto-correlation effectively resolves the arbitrary initialization of the population.

$$:{X}_{n+1}=left{begin{array}{l}modleft[mu:{X}_{n}left(1-{X}_{n}right)+frac{left(4-mu:right){X}_{n}}{2}right],:{X}_{n}<0.5\:mod{:left[beta:r{X}_{n}left(1-{X}_{n}right)+frac{left(4-mu:right)left(1-{X}_{n}right)}{2}right],:X}_{n}ge:0.5end{array}right.:$$

(14)

(:n) represents the location of arbitrarily generated HHO, and (:{X}_{n+1}) indicates the location of (:n)th HHO once chaotic mapping is composited.

Population hierarchies improve search strategy

This approach has a type of population hierarchy, choosing the first five optimum locations in every iteration process rather than the one optimum location accepted initially. This increases the interaction among populations and effectively decreases the risky model falling into the local optimal.

$$:{X}_{rabbit}=frac{sum:_{i=a}^{e}fleft({X}_{ir}left(tright)right)}{sum:_{j=a}^{e}fleft({X}_{jr}left(tright)right)}cdot:Xleft(tright):$$

(15)

(:a,b,) (:c,d), and (:e) are the top 3 optimum locations regarding fitness value, (:{X}_{rabbit}) specifies the optimum location chosen in every iteration, (:t) represents iteration counts, and (:fleft({X}_{ir}left(tright)right)) indicates the fitness value of the optimum location in the 10-th iteration.

Enhancement of reducing escape energy approach

It is motivated by the fractional order predator-prey dynamical technique with housings; an advanced model is projected for escaping the energy-reducing approach, which presents a contraction function in the energy-reducing equations to prevent the model from getting stuck in local optima by ensuring the escaping energy (:E) remains lower in mid and late iterations.

$$:{E}_{1}=2cdot:randcdot:{e}^{-left(frac{pi:}{2}cdot:left(frac{c}{T}right)right)}:$$

(16)

$$:E=2{E}_{0}cdot:{E}_{1}$$

(17)

The IHHO model initiates an FF to obtain heightened classification performance. It summarizes progressive numbers to characterize the enriched outcome of the candidate solutions. In this paper, the minimization of the classification rate of error is imitated as the FF, as provided in Eq. (18).

$$begin{aligned} fitnessleft( {x_{i} } right) & = ClassifierErrorRateleft( {x_{i} } right) \ & = frac{{no:of:misclassified:samples}}{{Total:no:of:samples}} times 100 \ end{aligned}$$

(18)



Source link

AI Insights

In the ever-changing artificial intelligence (AI) world, there is a place that is gaining an unrival..

Published

on


In the ever-changing artificial intelligence (AI) world, there is a place that is gaining an unrivaled status as an AI-based language-specific service. DeepL started in Germany in 2017 and now has 200,000 companies around the world as customers.

DeepL Chief Revenue Officer David Parry Jones, whom Mail Business recently met via video, is in charge of all customer management and support.

DeepL is focusing on securing customers by introducing a large number of services tailored to their needs, such as launching “Deep L for Enterprise,” a corporate product, and “Deep L Voice,” a voice translation solution, last year.

“We are focusing on translators, which are key products, and DeepL Voice is gaining popularity as it is installed in the Teams environment,” Pari-Jones CRO said. “We are also considering installing it on Zoom, a video conference platform.”

DeepL’s voice translation solution is currently integrated into Microsoft’s Teams. If participants in the meeting using Teams speak their own language, other participants can check subtitles that are translated in real-time. As the global video conference market accounts for nearly 90% of Zoom and MS Teams, if DeepL solutions are introduced through Zoom, the language barrier in video conferences will now disappear.

DeepL solutions are all focused on saving time and resources going into translation and delivering accurate results. “According to a study commissioned by Forrester Research last year, companies’ internal document translation time was reduced by 90% when using DeepL solutions,” Parry Jones CRO said, explaining that it is playing a role in breaking down language barriers and strengthening efficiency.

The Asian market, including Korea, a non-English speaking country, is considered a key market for DeepL. CEO Yarek Kutilovsky also visits Korea almost every year and meets with domestic customers.

“The Asia-Pacific region and Japan are behind DeepL’s rapid growth,” said CRO Pari-Jones. In translation services, the region accounts for 45% of sales, he said. “In particular, Japan is the second largest market, and Korea is closely following it.” He explains that Korea and Japan have similar levels of English proficiency, and there are many large corporations that are active in multiple countries, so there is a high demand for high-quality translations.

In Japan, Daiwa Securities is using DeepL solutions in the process of disclosing performance-related data to the world, and Fujifilm and NEC are also representative customers of DeepL. In Korea, Yanolja, Lotte Innovate, and Lightning Market are using DeepL.

DeepL currently only has branches in Japan among Asian countries, and the Korean branch is also considering establishing it, although the exact timing has not been set.

“DeepL continues to improve translation quality and invest at the same time for growth in Korea,” said CRO Pari-Jones. “We need a local team for growth.” We can’t promise you the exact schedule, but (the Korean branch) will be a natural development,” he said.

Meanwhile, as Generative AI services such as ChatGPT become more common, these services are also not the main function, but they also perform compliance levels of translation, threatening translators.

DeepL also sees them as competitors and competes. “DeepL is a translation company, so the difference is that it strives to provide accuracy or innovative language services,” Pari-Jones CRO said. “When comparing translation quality, the gap has narrowed slightly with ChatGPT.” We will continue to improve quality while testing regularly,” he said.

[Reporter Jeong Hojun]



Source link

Continue Reading

AI Insights

There is No Such Thing as Artificial Intelligence – Nathan Beacom

Published

on


One man tried to kill a cop with a butcher knife, because OpenAI killed his lover. A 29-year-old mother became violent toward her husband when he suggested that her relationship with ChatGPT was not real. A 41-year-old now-single mom split with her husband after he became consumed with chatbot communication, developing bizarre paranoia and conspiracy theories.

These stories, reported by the New York Times and Rolling Stone, represent the frightening, far end of the spectrum of chatbot-induced madness. How many people, we might wonder, are quietly losing their minds because they’ve turned to chatbots as a salve for loneliness or frustrated romantic desire?



Source link

Continue Reading

AI Insights

Apple Supplier Lens Tech Said to Price $607 Million Hong Kong Listing at Top of Range

Published

on




Apple Inc. supplier Lens Technology Co. has raised HK$4.8 billion ($607 million) after pricing its Hong Kong listing at the top of the marketed range, according to people familiar with the matter.



Source link

Continue Reading

Trending