An adaptive hybrid african vultures‑aquila optimizer with Xgb‑Tree algorithm for fake news detection

Online platforms and social networking have increased in the contemporary years. They are now a major news source worldwide, leading to the online proliferation of Fake News (FNs). These FNs are alarming because they fundamentally reshape public opinion, which may cause customers to leave these online platforms, threaten‑ ing the reputations of several organizations and industries. This rapid dissemination of FNs makes it imperative for automated systems to detect them, encouraging many researchers to propose various systems to classify news articles and detect FNs auto‑ matically. In this paper, a Fake News Detection (FND) methodology is presented based on an effective IBAVO‑AO algorithm, which stands for hybridization of African Vultures Optimization (AVO) and Aquila Optimization (AO) algorithms, with an extreme gradient boosting Tree (Xgb‑Tree) classifier. The suggested methodology involves three main phases: Initially, the unstructured FNs dataset is analyzed, and the essential features are extracted by tokenizing, encoding, and padding the input news words into a sequence of integers utilizing the GLOVE approach. Then, the extracted features are filtered using the effective Relief algorithm to select only the appropriate ones. Finally, the recovered features are used to classify the news items using the suggested IBAVO‑AO algorithm based on the Xgb‑Tree classifier. Hence, the suggested methodology is distinguished from prior models in that it performs automatic data pre‑processing, optimization, and classification tasks. The proposed methodology is carried out on the ISOT‑FNs dataset, containing more than 44 thousand multiple news articles divided into truthful and fake. We validated the proposed methodology’s reliability by examining numer‑ ous evaluation metrics involving accuracy, fitness values, the number of selected features, Kappa, Precision, Recall, F1‑score, Specificity, Sensitivity, ROC_AUC, and MCC. Then, the proposed methodology is compared against the most common meta‑ heuristic optimization algorithms utilizing the ISOT‑FNs. The experimental results reveal that the suggested methodology achieved optimal classification accuracy and F1‑score and successfully categorized more than 92.5% of news articles compared to its peers. This study will assist researchers in expanding their understanding of meta‑ heuristic optimization algorithms applications for FND.


Introduction
FND techniques have been getting extra attention since the circulation of disinformation has increased on the Internet, which has become a concern of the modern community [1].Generally, the concept of FNs has been around for a while.This problem existed before the growth of the Internet.Many publishers utilize misinformation to promote their interests [2].Many publishers publish FNs through convenient print media news and online platforms.Online platforms play an essential role in disseminating FNs in the community; these online platforms, such as online newspapers and social media, provide users access to various publications in one session to provide greater ease and speed than printed news media.In addition, the nature of social networks suggests an accessible platform for the fast dissemination of information in real-time; even with the reliability of this information, it has caused severe information credibility problems [3].
Not only do FNs negatively affect individuals, but they devastate the community as a whole over time.For example, FNs went viral on Facebook in the US 2016 presidential election instead of the more popular and trusted traditional news sources [4], revealing that readers may pay more attention to FNs than truthful news.Social media users who participate in spreading disinformation can have many motivations for spreading such information online, such as manipulation, political agendas, and influence.Still, while many of these users are genuine, those spreading disinformation may or may not be genuine users [5].Because social media profiles are inexpensive and uncomplicated, many people have created social media profiles for malicious tasks.If a computer algorithm manages social media profiles, it will be used as a social bot [4].These social bots can interact with individuals via social media and automatically produce and publish content online, making it significantly challenging for individuals to recognize such manipulated content [6].
Therefore, it isn't easy to validate online content using manual methods because, in recent years, a large amount of online content has been created and published online.Moreover, many researchers emphasized that automated and computerized FND methods should be used and are necessary [7].FND systems have generally been divided into "news content" and "social context" classes using their information sources.The first class is "news content" techniques, which attempt to validate news content and utilize attributes such as body text, title, and more metadata to recognize FNs.These techniques are called "content-driven" techniques [8].The Second category is social context techniques that focus on social attributes such as users' interactions in social media with specific news (liking it or sharing it on Facebook, retweeting it on Twitter).These techniques are referred to as "social-driven" techniques [8].
Deep Learning (DL) and machine learning methods have been employed in different regions [9][10][11] and have recently been used to tackle FND issues effectively and efficiently [12].The leading cause for effective outcomes utilizing DL techniques is the large data volume and high dimensionality of the data for FNs.Today's scenario is a fast and large-scale growth of social media, and people are using social media to view the latest updates.Thus, social media platforms such as WhatsApp, Twitter, Facebook, and YouTube struggle to detect FNs from many user posts.There is a potential danger of publishing and disseminating such FNs via social media Platforms [13].Many challenges must be considered when working in these areas, including selecting the most appropriate attributes, high-dimensionality data, heterogeneity, and choosing the most appropriate DL technique [14].[15] proposed a DL method based on an automated detector via a three-level hierarchical focus network for fast and accurate FND.[16] proposed deep Convolutional Neural Networks (CNNs) for detecting FNs.[14] presented a learning model based on linguistic features to detect FNs.[17] presented a method for FND using a hybrid neural network structure, integrating the power of Long Short-Term Memory (LSTM) and CNNs.[13] presented several attributes-oriented methods for the automated detection of FNs on social media employing DL. [18] presented three DL-based models intended to classify and detect FNs.[19] presented a method for FND employing a geometric DL. [20] introduced a neural network method to accurately forecast the stance between a given pair of headlines and the text of the article.[21] introduced several methods for FND based on the relationship between the headlines and the body of the articles.Their methods are primarily based on Bidirectional-LSTM, CNN, and LSTM.
Due to their effective performance in addressing many optimization problems, metaheuristic algorithms have attracted much attention recently.Therefore, MHA is an efficient solution-finding method to detect FNs on social media.[22] introduced the issue of detecting FNs as an optimization problem.This study proposes two meta-heuristic algorithms, Grey Wolf Optimization (GWO) and Salp Swarm Optimization (SSO), for tackling the FND issue.The proposed FND approach is initialized through a pre-processing phase and then utilizes GWO and SSO to handle the FND issue.The suggested approach was verified utilizing three real-world FNs datasets.The experimental outcomes show that the GWO optimization algorithm achieved optimal results in different performance metrics than the SSO optimization algorithm and other meta-heuristic algorithms.[23] improved their study by proposing a new method that integrated MHA and text mining to discover FNs via online social media.Modified variants of GWO and SWO optimization algorithms based on nonlinear decreasing coefficient and oscillating inertia weight are used for the FND issue.The evaluation measures of the suggested approaches are verified on different datasets.The empirical outcomes revealed that the proposed new approaches exceeded other approaches in real-world FNs datasets.[24] introduced a new method for identifying FNs articles using the WOA-Xgb-Tree technique and content-driven attributes.The suggested model can be implemented in several scenarios for classifying news articles.The proposed model has two phases: first, the necessary attributes are identified and investigated.Then, the Xgb-Tree optimizer tuned by the Whale Optimization Algorithm (WOA) classifies the news articles using the specified attributes.In their empirical results, They considered F1-score and classification accuracy as the basis of their investigations.Then, they compared the results of their proposed system to various modern classification techniques using a dataset that has collected more than 40,000 news articles recently.The empirical outcomes reveal that the suggested system obtained a reliable F1-score rate and efficiently classified more than 91 percent of the articles.

Motivations
This paper presents a framework relying on the IBAVO-AO algorithm to tackle the issue of FND.The proposed IBAVO-AO is a hybrid AVO-AO optimizer with an Xgb-Tree classifier.The primary stages of the suggested methodology are as follows: Firstly, the collected unstructured data is converted into structured data for usage in the classification process, known as data pre-processing.In this stage, beneficial features are extracted by removing superfluous words and unnecessary special symbols, stemming from altering words into root words, tokenizing the resulting data into a bag of words, and finally encoding and padding words into sequence vectors of numerical values using Global Vectors (GLOVE) [25,26], which is a count-based approach for pre-training and relies on terms or vectors from co-occurrence data.After that, the extracted features are filtered using an efficient Relief algorithm to determine only the associated features and provide the final classification dataset.Using the Relief algorithm aims to enhance the ability to explore the best outcomes discovered inside the solution space.In the final stage, the classification process utilizes the IBAVO-AO algorithm based on the Xgb-Tree classifier with high detection performance.The effectiveness of the suggested methodology is assessed by employing a variety of evaluation metrics and applying them to the ISOT-FNs dataset that includes more than 44 thousand news articles.After the suggested methodology has been evaluated and compared with state-of-the-art optimization techniques [27,28], the results indicate that the presented methodology produces high classification accuracy.It is advised to use it in the FND problem.

Contributions
This paper offers an FND methodology based on the IBAVO-AO algorithm with Xgb-Tree classifier; its fundamental contributions can be clarified in the following points: • Pre-process the FNs data to extract the necessary features.
• For improving and reducing the initial search space exploration capacity and enhancing the acquired optimal outcomes, the proposed IBAVO-AO algorithm embeds a Relief algorithm with the hybridization of AVO and AO algorithms.This embedding enhances the algorithm's performance by producing a new population that maintains the fundamental structure but has more appropriate positions.• Filter and determine only the most appropriate features for predictive modeling using the Relief algorithm.• Classify the news items utilizing the IBAVO-AO with the Xgb-Tree classifier.
• Assess the proposed methodology against state-of-the-art optimization algorithms using a variety of evaluation metrics involving accuracy, fitness values, the number of selected features, Kappa, Precision, Recall, F1-score, Specificity, Sensitivity, ROC_ AUC, and MCC toward the ISOT-FNs dataset.• The proposed methodology outcomes achieve high classification accuracy and a positive impact compared to its peers.

Structure
The remainder of the paper is formulated as follows.Section "Literature review" presents the related work and literature review.The proposed methodology and its components for FND are presented in Sect."The proposed IBAVO-AO algorithm for FND".Section "Experimental results and analysis" shows the numerical results and comparisons.Finally, conclusions and future work are drawn in Sect."Conclusion and future work".

Literature review
The primary purpose of fact-checking is to use new technologies to recognize unreliable and manipulated news content on the Internet.It is an attractive major topic within specific streams of information and library science [29].As a result, many researchers are trying to address the issues of FNs in different areas, especially online news.This section will survey the various methods utilized to discover FNs on online platforms, and we will briefly mention their results and advantages.[30] has proposed a new model for detecting real and fake stories.They used linguistic attributes like special characters, emoticon symbols, negative/positive words, and hashtags to categorize news stories.[31] suggests a system for detecting information sequences in a Twitter OP.Within their work, patterns analysis methods were implemented, allowing them to classify original and FNs.[32] proposed a graph kernel-based Support vector Machine (SVM) method that learns high-order distribution patterns to detect FNs.[33] proposed a novel model that uses a Recurrent Neural Network (RNN) to identify FN articles utilizing linguistic attributes accessed from a collection of user comments.
[34] introduced a novel system to identify authentic news articles.They utilized inevitable connections among conversation parts to identify trustworthy news stories.In another system, [35] analyzed the same user features on the social platform Sina Weibo, the most popular Chinese microblogging site.[36] proposed an approach to identify sarcastic tweets and product reviews automatically.They have used generic attributes based on baseline and lexicon features.The features such as character n-grams, word n-grams, and word skip-grams are extracted and integrated with lexicon properties.Then, they categorized these features utilizing different methods, such as ensemble classifiers, Logistic Regression (LR), Decision Tree (DT), and Random Forest (RF).
[37] studied the value of various attributes in identifying and categorizing sarcastic and ironic reviews on different types of products.Firstly, they elicited attributes utilizing lexicon-based features and Bag-of-Words.Then, they used these elicited attributes on various Machine Learning (ML) classifiers such as LR, SVM, DT, and RF.[38] suggested a novel system to recognize truthful news.They integrated the different user, linguistic, structural, and temporal features to categorize FNs.[39] proposed a method to address the problem of FNs based on DL.The suggested method contains three stages: text encoding, feature extraction, and classification.The text encoding phase is performed on the entered news words utilizing GLOVE to represent the words.The encoded words at a given word length are then included to be enrolled in the suggested DL methods.The suggested DL methods include both automated feature extraction and classification capabilities.Moreover, this search presents four DL methods containing CNNs and Concatenated CNNs, Gated Recurrent Units, and LSTM to obtain an optimum method before the problem of FNs exceeds previous studies.The suggested DL methods are implemented on FNC and FNs datasets supported by Kaggle.The proposed Concatenated CNNs method achieved a classification accuracy of 99.6% and trained faster than others.
Table 1 presents the state-of-the-art papers implemented in FND, including the dataset, model description, limitations, advantages, and outcomes.According to the outcomes of Table 1, there are many issues still open in this area, which can be summarized in the following points: • The limited suitable quality-labeled benchmark datasets.
• Little studies have been implemented on regional languages.• Lack of a comprehensive standardized data set on FNs.

Advantages
Outcomes [12] ISOT and FA-KES The study introduced a novel hybrid DL approach integrating RNNs and CNNs for FND • Very few studies have been done on detecting FNs as a multi-category classification problem.• The DL algorithms have poor re-usability and transfer learning capability.
• The accuracy of classification is not effective in many studies.
• The performance of many of these studies in detecting FNs is still insufficient.
These issues encouraged us to introduce a novel approach to classifying news articles automatically utilizing content-based attributes and useful linguistics and the proposed hybrid algorithm based on AVO and AO with the Xgb-Tree algorithm for detecting FNs with high performance.

Data pre-processing
During this stage, the unstructured data gathered from the suggested FNs dataset (see Sect. "Dataset description") is transformed into structured data for classification.The two primary methods of this stage are to extract features from the FNs dataset, then filter the consequent features and elect only the pertinent ones.These methods are covered in the subsequent subsections.

Feature extraction
This method is carried out as follows: first, extraneous words and special symbols that are not required, such as digits, stop-words, words with only one letter, commas, hashtags, punctuation marks, etc., are eliminated from the unstructured data for an FNs dataset.Then, the necessary words are adjusted into root words, called stemming, by ignoring suffixes, affixes, inserts, and a mix of starting and ending on derived words.After correcting misspellings or abbreviations, the remaining words are changed to lowercase to handle a unified form.
The resultant data are tokenized by segregating them into a bag of words (small tokens) to obtain the words that have value in the created matrix for employment in the classification process.Additionally, the data is encoded into sequence vectors using the GLOVE method [25] for word representation, which turns tokens into a sequence of integers.Since the encoded data used are of varying lengths, each sequence vector is padded to ensure that all sequences have identical lengths; this is done by padding zero at the start of each sequence until each one equals the maximum length specified for each padded vector, which is set to 1000.Labels are also encoded, with positive labels encoded as one and negative labels encoded as zero.

Feature filtration
The sheer number of features is considered one of the significant challenges in the data pre-processing stage.Processing time and computing effort are often increased while dealing with these numerous features.Also, it could hurt classification performance.Thus, there is a need to introduce an effective method for filtering the features and picking appropriate ones.This paper suggests a straightforward and quick filtering method called the Relief algorithm [48,49], which is suggested to identify related features.This method focuses only on pertinent features and reduces the initial search space by locating features with comparable values for identical close samples and significant for the difference between dissimilar samples.According to the features' weighted ranking, the algorithm works as follows: First, it distinguishes between Near-Hit samples related to congruent class samples and Near-Miss samples related to mismatched class samples.The weight of the feature is then evaluated based on the Near-Hit and Near-Miss values to assess the suitability of the classification process.The features are then ranked according to their weights from the biggest to the least.The following equation can be employed to evaluate the feature weight W A : where W A denotes the weight of the feature, N is the sample number, and X j A means the feature value A of data X j .NH(X j ) and NM(X j ) indicate the nearest data points to X j related to the identical and the distinct class, respectively.The Relief algorithm holds significance in feature filtration and is suitable for this problem due to the following key reasons: • Robustness with noisy data: The Relief algorithm is known for its robustness in handling noisy data.If the presented datasets contain noise or outliers, Relief can perform well despite these challenges.It evaluates feature importance by considering the proximity between instances, which helps mitigate the impact of noise or irrelevant features.• Capability to identify relevant features: The Relief algorithm is designed to identify the most relevant features by considering their contribution.• Balancing feature relevance: The Relief algorithm considers feature relevance and redundancy factors.It helps identify a subset of features that contribute most substantially to the model's performance by differentiating between those that may be redundant and those most relevant to the target variable.• Bias-Free feature selection: the Relief algorithm is less prone to exhibit bias in feature selection since it does not make any assumptions about any specific data distribution.This feature guarantees a more impartial assessment of feature significance.• Efficiency in handling high-dimensional data: The Relief algorithm performs well and is appropriate for datasets with substantial features, handling high-dimensional data efficiently.Compared to certain feature selection methods, Relief could be preferred as it tends to perform well without suffering from the curse of dimensionality.• Handling diverse data types: The Relief algorithm is adaptable to a wide range of dataset types since it can handle continuous and categorical features.• Simple implementation and interpretability: The Relief algorithm is comparatively simple and easy to implement and comprehend, and its outcomes are frequently interpretable.Its usefulness is improved by its simplicity, particularly when the interpretability of feature selection is important.• Previous success or familiarity: The choice of the Relief algorithm might also stem from its success or prevalence in similar studies or datasets.It has been successfully applied in various problem domains, including healthcare, finance, and bioinformatics. ( Overall, the Relief algorithm's significance lies in its ability to efficiently and rapidly filter out unnecessary or redundant features, leading to improved model performance and interpretability even on noisy and varied datasets.

The proposed AVO algorithm
An efficient nature-inspired meta-heuristic optimization algorithm termed AVO algorithm [50] is presented in this paper for modeling and imitating the natural behaviors of vultures in Africa concerning living and nutrition behavior.This algorithm is set up dependent on basic conceptions related to vultures, as follows: Initially, the AVO algorithm assumes that the population size of African vultures consists of N vultures, which vary depending on the problem being tackled.After that, the fitness function value is computed for all solutions of the African vultures' initial population, allowing the vultures' population to be tangibly split into three sets; the first set comprises the best solution, which is a vulture that is stronger than all other vultures, the second set contains the second-best solution, which is the weaker vulture than the first set, and the last set has the remaining weakest African vultures.These three sets can formulate the most significant natural function of vultures.Each set has a unique incapability to obtain and consume food.Further, the advantages and disadvantages of vultures may be reflected in the fitness function's value of the solution.As a result, the two best solutions characterize the best and strongest vultures, whereas the worst solution represents the weakest and most starved vultures.In general, the vultures attempt to retain a safe range from the worst while attempting to get close to the best vultures.
According to the conceptions mentioned above, the proposed AVO algorithm can be formulated into four essential steps to model the behavior of various vultures.These steps are depicted in the next few subsections.
Population splitting step This step aims to divide the initial population into sets by evaluating the fitness function of their solutions.The first set includes the best solution as the first set's best vulture and the second-best solution is chosen as the second set's best vulture.The residual solutions are in the final set.The population should be re-formulated for each iteration because the solutions always try to come as near as possible to the best and second-best solutions, as follows: where BestVulture g 1 and SecondBestVulture g 2 denote the first set's best vulture and the second-best vulture in the second set at the g th iteration, respectively.pr g is the likeli- hood of choosing the most suitable solution for each set at the g th iteration, which is defined using the Roulette wheel procedure illustrated in Eq. (3).L 1 and L 2 are two ran- dom parameters within the range [0, 1]. (2) Vultures' starvation level step In this step, the amount of starvation is measured for the vultures, which can also be used mathematically to model the processes of exploration, exploitation, and transformation among them.The vultures can fly farther and have more capacity to look for food when they are not starving.On the other hand, Vultures cannot fly long distances for food and might turn hostile when starving.The starvation level ( F g i ) of the i th vulture at the g th iteration can be expressed as follows: The variable F g i indicates the vulture's transition from exploration to exploitation, which implies that the vultures are full.A rand is an arbitrary number between 0 and 1, z implies an arbitrary value within the range [−1, 1] , g i signifies the present iteration's number, and G max indicates the maximum iteration's number.The t value is computed by Eq. ( 5) to improve the effectiveness in tackling complex optimization problems and avoid falling into a local optimum.h is an arbitrary value within the range [−2, 2] .The predetermined constant parameter w determines the likelihood of performing the exploration process; the likelihood of exploration increases as its value rises.As its value drops, the likelihood of exploration decreases.
According to Eq. ( 4), the value of F g i progressively reduces with the increasing number of iterations.Therefore, the next step can be defined in the proposed AVO algorithm as follows: AVO's exploration step During the exploration step, the vultures are distinguished by their high capacity and optical ability to seek suitable food.Vultures are compelled to fly long distances for extended periods and inspect various random sites for food.Hence, the exploration step utilizes two distinct techniques.A predefined parameter P 1 and a ran- dom value rand P 1 are employed to pick one of these techniques with values in the range [0, 1].Notice that the starvation level |F g i | in the exploration step is more major than or equal to 1.The exploration techniques can be explained as follows: where is the vulture's next updated position at the next (g + 1) th iteration, R g is the chosen best vulture in the present iteration g, which is specified through Eq. ( 2), D g i is calculated using Eq. ( 8), F g i is the starvation level of the i th vulture at the g th iteration, (4) estimated by Eq. ( 4).rand is a random value amidst zero and one; to keep food safe from other vultures and to provide a high arbitrary coefficient at the search environment scale, the vultures move randomly.UB indicates the variables' upper limit, LB presents the variables' lower limit, and X g i is the present position at the g th iteration.
AVO's exploitation step In the AVO's exploitation step, the value of |F g i | is smaller than 1.The exploitation step consists of two internal sub-steps, where the effectiveness of the proposed AVO algorithm is assessed.Each of these sub-steps has two distinct techniques.Two predetermined parameters with values between 0 and 1 are utilized to specify the appropriate technique in each internal sub-step: P 2 for the first sub-step and P 3 for the second sub-step.The following is an explication of these two internal sub-steps.

First exploitation sub-step:
This sub-step is executed when the |F g i | value is smaller than 1 and greater than or equal to 0.5, which utilizes two distinct techniques.A predefined parameter P 2 and a random value rand P 2 , with values ranging from [0, 1], are employed to decide which of these two techniques is selected.
The first technique of this sub-step is known as siege-fight, in which the vultures have enough power and are moderately satiated.Because vultures gather around one specific food source, the stronger and healthful vultures attempt not to exchange food with others.In contrast, the more powerless vultures attempt to steal food from the healthful vultures by swarming close to them and starting little fights.On the other hand, the second technique is referred to as rotational-flight; it models and forms a spiral motion between one of the best vultures and the remaining.The techniques of the first exploitation sub-step can be illustrated as follows: where X g+1 i denotes the vulture's next updated position at the following (g + 1) th iteration, D g i is derived using Eq. ( 8), F g i indicates the degree of starvation for the i th vulture at the g th iteration as determined by Eq. ( 4), and rand is a random value between 0 and 1 to provide a high arbitrary coefficient.d g t is the distance between the vulture and one of the best two vultures, which is estimated by Eq. ( 10), R g means the preferred best vulture in the present g th iteration, which is set via Eq.( 2), S g 1 and S g 2 (9) are estimated utilizing Eq. ( 11) and ( 12) respectively, and X g i represents the present position at the g th iteration.
2. Second exploitation sub-step: when the value of |F g i | is less than 0.5, this sub-step is implemented, in which numerous sieges and violent fights are performed among diverse species of vultures that have congregated around the food source.Two various techniques are used in this sub-step.To determine which of these two techniques to select, a predetermined parameter P 3 and a random value rand P 3 , with val- ues ranging from 0 to 1, are created.
Congregate vultures around the food source is the name of the first technique of this sub-step, as diverse species of vultures are hungry so that they may attract and compete near a single food supply.Furthermore, the second technique is termed an aggressive siege-fight.The vultures become more offensive and attempt to scavenge the remaining food from the healthful vultures by flocking toward them in various directions.In contrast, the healthful vultures weaken and lose the power to resist the other vultures.The techniques of the second exploitation sub-step can be depicted as follows: where X g+1 i signifies the vulture's next updated position at the following (g + 1) th iteration, which reflects the congregation of vultures.A g 1 and A g 2 are assessed by using Eq. ( 14) and ( 15) respectively, R g means the choice best vulture at the present g th iteration, which is defined via Eq.( 2), d g t stands for the distance between the vulture and one of the best two vultures, which is estimated by Eq. ( 10), F g i indicates the degree of starvation for the i th vulture at the g th iteration that computed by Eq. ( 4), and Levy d is the function of levy flight distribution acquired by Eq. ( 16) to improve the efficiency of the AVO algorithm.The best vulture in the first set and the secondbest vulture in the second set at the present g th iteration are denoted by BestVulture g 1 and SecondBestVulture g 2 respectively, while the present position at the g th iteration is represented by X g i .d is the dimensional space, µ and ν are arbitrary values evenly distributed throughout the range [0, 1], and σ is specified by Eq. ( 16), where β = 1.5 is a constant number.(13) .
Pseudo-code of the proposed AVO algorithm Following clarifying the critical steps of the suggested AVO algorithm illustrated above and presenting the techniques that are recommended for mimicking the natural behaviors of African vultures in living and feeding, the pseudo-code defining the proposed AVO algorithm is provided in Algorithm 1.Moreover, a flowchart of the AVO algorithm is shown in Fig. 1 to highlight its main steps.

The proposed AO algorithm
In this subsection, the prey-hunting behavior of Aquila is simulated by introducing an efficacious meta-heuristic optimization algorithm inspired by nature, dubbed the AO algorithm [51].Due to its bravery, agility, and speed, depending on steady feet and sharp talons when hunting various animals, including badgers, squirrels, and rabbits, the Aquila is the most famous prey bird [52].Aquila relies on four hunting techniques, which can be summed up mathematically in two crucial steps: exploration and exploitation.The appropriate step can be chosen from these two steps in the proposed AO algorithm according to the present iteration's number, g i , and the maximum iteration's number, G max , as follows: The following subsections illustrate these steps of the proposed AO algorithm.
AO's exploration step The exploration step includes two distinct techniques.Extensive exploration is the name of the first technique, where the Aquila flies far above the land in search of suitable prey.The Aquila begins a long, low-angled glide with growing speed as it approaches the optimal region for prey.It then extends its wings and tail and lowers vertically toward the prey.On the other hand, the second technique is restricted exploration, in which the Aquila carefully inspects the chosen area of the prey from a high altitude, whether the prey is in flight or a running state.The Aquila then spirals around the chosen prey and goes up low off the ground to get ready to catch the prey.A random value rand, ranging from [0, 1], determines which of these two techniques to pick.
For improving the exploration's efficiency, the exploration step can be mathematically stated when g i is smaller than or equal to 2  3 • G max .The techniques mentioned above of the AO's exploration step can be represented as follows: where X g+1 i denotes the aquila's next updated position at the subsequent (g + 1) th itera- tion, X g Best indicates the present best position found during searching at g th iteration.g i means the current iteration, while G max means the maximum allowed iterations' num- ber, the phrase 1 − g i G max is used to dominate the extended exploration throughout the set of iterations.X g Mean is the mean value of the present positions at g th iteration, which is evaluated through Eq. (19).The number of permitted positions is N, and d is the problem's dimension size.Levy d is the levy flight distribution function, emanated using Eq. ( 16).X g τ is a randomly chosen Aquila's position.The twisting form in the search is represented by y g and ζ g , which are evaluated using Eq. ( 20) and (21), respectively.r 1 indicates (17) the number of search rotations ranges from 1 to 20, and U = 0.00565 .D 1 is integer numbers from 1 to d, and ω = 0.005.
AO's exploitation step Two diverse techniques are used in the exploitation step.The first technique is dubbed extensive exploitation.In this technique, the Aquila lands on the ground after exactly locating and exploiting the prey region and slowly approaches it for catching.This technique suits slow-moving prey or prey that lacks an escape response.
Restricted exploitation is the name of the second technique used in the exploitation step, in which the Aquila moves on the ground as it nears and attacks its prey at the last location by following its random motions.A random number rand, with a value between [0, 1], is employed to choose between these two techniques.Mathematically, in the exploitation step, when g i is greater than 2 3 • G max , X i is mod- ified for enhancing the exploitation's performance.The aforementioned exploitation techniques of the AO can be illustrated as follows: where X g+1 i is the aquila's next updated position at the following (g + 1) th iteration, X g Best means the current best position found during the search at g th iteration.X g Mean denotes the mean value of the present positions at g th iteration, and can be assessed through Eq. (19).The exploitation step's adjustment parameters, α , and δ , are given to (0.1).UB and LB indicate the upper and lower limits of the search space, respectively.The search strategy is balanced using the quality function value QF g , which is calculated using Eq. ( 23).Aquila's arbitrary motions while pursuing its prey are reflected in Q g 1 by Eq. ( 24).Aquila's flying slope when tracking its prey is represented by Q g 2 , which decreases in value from 2 to 0 and is determined using equation Eq.( 24).X g i is the present position at the g th itera- tion, Levy d is the function of levy flight distribution, defined using Eq. ( 16), g i means the current iteration, while G max represents the maximum allowed iterations' number.
Pseudo-code of the proposed AO algorithm After introducing the steps mentioned above of the AO algorithm, exploration and exploitation, and showing the four techniques suggested to imitate Aquila's hunting behavior, the pseudo-code of the proposed AO algorithm is presented in Algorithm 2. Additionally, Fig. 2 includes a flowchart of the AO algorithm to show its main steps.( 22) .

Adjustment of bound-constraint
This paper presents a bound-constrained adjustment method for re-positioning impractical decision variables beyond the search space's scope during position improvement employing the above-mentioned meta-heuristic optimization algorithms (AVO and AO algorithms).Using the random method that adjusts decision variable values outside the permissible limits with randomly generated ones inside those limits is recommended.This method can be mathematically stated as follows: Where represents the value of the appropriate decision variable, X i,d denotes the infeasible value that is beyond the variable's limits, X LB d and X UB d depict the lower and (25) upper limits, respectively, and rand(0, 1) is a random number falling within the range [0, 1].

Xgb-Tree classification algorithm
The Xgb-Tree [53] is a developed algorithm for the gradient boosting framework [54][55][56], which can classify sample instances into a specific class.This algorithm utilizes integrating methods and additional training procedures to integrate many weak learners to create a powerful learner.The Xgb-Tree algorithm's core concept is to boost the gradient tree by consecutively producing DTs.According to complementary models from prior iterations, boosting decreases errors and enhances classification performance.The Xgb-Tree's objective function consists of a training loss part, which Where T L stands for the whole leaves' number on the tree, and ϕ j signifies a vector value representing leaves' scores.

The proposed IBAVO-AO algorithm
Since FNs are purposefully designed to provide false information, detecting it can be challenging.This paper suggests an effective IBAVO-AO algorithm to specify FNs by combining the AVO algorithm and the AO algorithm, leading to more accurate findings.In our proposed IBAVO-AO algorithm, we tried to solve the FND problem by hybridizing the natural processes of AVO and AO.In the proposed algorithm, the AVO algorithm creates solutions in their search space and tries to improve them.After that, the AO algorithm improves the solutions produced in the space of the AVO solutions through exploration and exploitation processes.The IBAVO-AO algorithm combines the AVO and AO algorithms through the following steps: • Firstly, within the specified search space, the population of AVO solutions with random values is initialized.the AVO algorithm handles The exploration and exploitation of the search space.The exploration step permits the algorithm to search for new areas of the search space, while the exploitation step concentrates on boosting the search around promising solutions.AVO algorithm simulates an African vulture's living and feeding behavior to improve solutions iteratively.• The vultures navigate the search space by adjusting their positions based on their current positions and a set of candidate solutions that have already been discovered.• Secondly, the AO algorithm is another optimization technique used in the IBAVO-AO algorithm.It operates on a set of the candidate solutions obtained by the AVO algorithm and improves them iteratively.• The AO algorithm aims to improve the produced AVO solutions and strike a balance between exploration and exploitation capabilities.This balance facilitates the effective exploration of the search space, preventing the occurrence of local optima and enhancing the convergence towards optimum solutions.• A set of new candidate solutions is created during the AO algorithm and combined with the population's preexisting solutions.After that, a comparison between the new candidate solutions and the original solutions is made, and chosen is performed based on the values of their objective functions.
Combining the AVO and AO algorithms involves creating a hybrid IBAVO-AO algorithm that leverages the strengths of each algorithm to improve overall optimization performance, convergence speed, and solution quality.The main advantages offered by the hybrid IBAVO-AO algorithm integrating the AVO and AO algorithms over using them separately are as follows: (26) • Enhanced exploration and exploitation: AVO and AO algorithms might be supreme in different aspects of exploration and exploitation.Combining them allows the hybrid IBAVO-AO algorithm to explore a broader solution space effectively.• Diversity in search: AVO and AO algorithms have different search mechanisms, enabling the hybrid IBAVO-AO algorithm to maintain a diverse population of solutions.This diversity can prevent premature convergence to suboptimal solutions.• Improved convergence: Leveraging the complementary strengths of AVO and AO algorithms, the hybrid IBAVO-AO algorithm can converge faster toward better solutions than using each algorithm separately.• Robustness: The hybrid IBAVO-AO algorithm enhances the robustness of the optimization process.It will be more resilient to getting stuck in local optima.
The proposed IBAVO-AO algorithm divides News items into two class labels-Fake and Truthful-meaning that the FND issue is drafted as a binary classification.The suggested IBAVO-AO algorithm's flowchart is depicted in Fig. 3.The proposed methodology for Fig. 3 Flowchart of the proposed IBAVO-AO algorithm FND includes the following steps: Initially, as described in Sect."Data pre-processing", the FNs dataset is pre-processed employing feature extraction and feature filtration methods.After that, the ultimate classification dataset is generated utilizing the pertinent features.Eventually, the proposed IBAVO-AVO algorithm is applied to the developed dataset, which updates the positions and determines the best values depending on the Xgb-Tree classification algorithm.

Experimental results and analysis
This section details the experimental results to evaluate the suggested FND methodology based on the IBAVO-AO algorithm with the Xgb-Tree classifier, describes the evaluation measures, and discusses the classification results.

Dataset description
The ISOT-FNs dataset [57] is an extensive collection comprising approximately 44,900 news articles.This dataset is bifurcated into two primary categories: truthful and FNs.The methodology employed in compiling this dataset is meticulous, involving a selection of news articles from various sources, each meticulously vetted for reliability.For sourcing truthful news, the dataset relies on articles from Reuters, a well-regarded international news organization known for its comprehensive and factbased reporting.In contrast, the FNs articles are sourced from websites identified as unreliable by reputable fact-checking entities such as Politifact.com and Wikipedia.While the ISOT-FNs dataset offers a valuable resource for studying the characteristics and spread of fake versus truthful news, it's essential to consider potential biases.The selection of sources, particularly for FNs, might reflect biases inherent in the criteria used by Politifact.com and Wikipedia.This could result in a dataset that may not fully represent the spectrum of FNs sources, especially those that are more subtle or sophisticated in their misinformation strategies.
Moreover, when comparing the ISOT-FNs dataset to other popular datasets in the field, such as the FNs Challenge (FNC-1) dataset or the Liar dataset, there are noticeable differences in size, source diversity, and categorization methodologies.For instance, the FNC-1 dataset focuses more on the stance detection between headline and body text, whereas the Liar dataset includes short statements and speeches labeled for truthfulness.These differences highlight the varying approaches in the field of FND and the importance of considering multiple datasets to gain a comprehensive understanding of the issue.

Experimental setup
FND is a complicated process, and the appropriate method requires various factors to identify manipulated news efficiently.That is the leading cause for integrating the IBAVO-AO optimization algorithm and the Xgb-Tree classification algorithm into the suggested methodology.Moreover, in contrast to the different methods presented, our suggested approach employed the Relief algorithm, which is explained in Sect."Feature filtration", to pick only the relevant features and identify FNs articles in less time and with lower processing cost by calculating the weights of each feature in the dataset and sorting them from biggest to smallest.Lastly, the features that have small weight are removed.Upon executing the Relief algorithm on the datasets, we discovered that the greatest weights were only associated with 50 features.For this reason, just these 50 important features were selected, while the remaining irrelevant features with minor weights were omitted.
This method focuses only on pertinent features and reduces the initial search space by locating features with comparable values for identical close samples and significant for the difference between dissimilar samples.
Thus, to adequately assess the performance of the proposed system, two sets of experiments were carried out on the utilized ISOT-FNs dataset.In the first part of the experiments, we conducted a detailed comparative analysis using the ISOT-FNs dataset.To provide a robust benchmarking framework, we selected a variety of well-established classification algorithms, each known for its unique strengths in the domain of FND.These include: • Decision Tree (DT) [58]: A simple yet powerful algorithm valued for its interpretability and ease of use in various classification problems.• K-nearest Neighbors (k-NN) [59]: This algorithm is effective in handling multi-class classification tasks and is known for its simplicity and efficacy.• Gaussian Naive Bayes (GNB) [60]: Chosen for its proficiency in managing highdimensional data, GNB applies a probabilistic approach to classification.• Support Vector Machine (SVM) [61]: Renowned for its robustness, especially in high-dimensional spaces, making it suitable for complex classification tasks.• Random Forest (RF) [62]: Selected for its high accuracy and efficiency, especially in large datasets, RF is a versatile and powerful ensemble method.• Multilayer Perceptron (MLP) [63]: A feedforward artificial neural network known for its ability to learn non-linear models and patterns in data.These algorithms were rigorously tested against our IBAVO-AO with Xgb-Tree classification on the same dataset, providing a comprehensive and balanced benchmarking environment.
Table 2 shows the significant parameters of the classification algorithms introduced in this paper.Secondly, a comprehensive comparative analysis between our proposed IBAVO-AO combined with the Xgb-Tree classification algorithms and a range of widely recognized meta-heuristic optimization algorithms.These algorithms were meticulously chosen for their relevance and popularity in optimization tasks.They include: • Binary African Vulture Optimization (BAVO) [50]: An optimization algorithm inspired by the foraging behavior of vultures, known for its efficiency in binary search spaces.• Binary Aquila Optimizer (BAO) [51]: This algorithm mimics the hunting strategy of Aquila eagles and is notable for its precision and speed.• Binary Sparrow Search Algorithm (BSSA) [64]: A novel algorithm based on the social behavior of sparrows, appreciated for its ability to explore and exploit the solution space.
• Binary Atom Search Optimization (BASO) [65]: Inspired by the laws of physics and molecular movement, known for its robustness in binary optimization problems.• Binary Henry Gas Solubility Optimization (BHGSO) [66]: This algorithm simulates the gas solubility process and is recognized for its adaptability in various optimization contexts.• Binary Harris Hawks Optimization (BHHO) [67]: Mimics the cooperative hunting strategy of Harris hawks, known for its effectiveness in complex optimization scenarios.• Binary Sailfish Optimizer (BSFO) [68]: Based on the predatory behavior of sailfish, this algorithm is praised for its swift convergence and flexibility.• Binary Bat Algorithm (BBA) [69]: Utilizes echolocation behavior of bats and is popular for its balance between exploration and exploitation.• Binary Grasshopper Optimization Algorithm (BGOA) [70]: Inspired by the swarming behavior of grasshoppers, it's efficient in finding global optima in complex landscapes.• Binary Artificial Bee Colony (BABC [71]): Mimics the foraging behavior of honey bees, well-regarded for its simplicity and effectiveness in binary domains.• Binary Particle Swarm Optimization (BPSO) [72]: Based on the social behavior of bird flocking, this algorithm is known for its efficiency and easy implementation.
These selected algorithms represent a diverse range of strategies in meta-heuristic optimization, ensuring a robust and comprehensive benchmarking against our proposed IBAVO-AO algorithm.The comparative study was conducted on the ISOT-FNs dataset, Table 2 The major parameters of the classification algorithms and the specific parameters employed for each algorithm in the comparison are detailed in Table 3.

Xgb
Python was used on a computing environment with a Dual Intel ® Xeon ® Gold 5115 2.4 GHz CPU and 128 GB of RAM on the Microsoft Windows Server 2019 operating system to run all experiments in this study.For a reliable comparison, the size of the population is estimated to be ten, and the maximum number of iterations is estimated to be 100 for all methods.Accordingly, the population size was set to 10, and the number of iterations was set to 100.Also, in this study, the new dataset is split into learning and testing after defining content-oriented attributes and creating a new dataset.Thus, 80%  of the data was utilized for learning, while 20% was utilized for evaluating the proposed system.Finally, a 10-fold cross-validation method is employed to reduce model error for learning and testing purposes.

Evaluation measures
In this study, the effectiveness of the suggested IBAVO-AO with the Xgb-Tree methodology must be assessed utilizing standard metrics to ensure that the empirical outcomes are statistically valuable.To that end, the primary evaluation metric employed was accuracy [73], which is the number of successful predictions divided by the total number of predictions.
Accuracy is expressed as in Eq. ( 27): Where True Positive (T P ) is the percentage of FNs that were successfully classified utiliz- ing the proposed system, True Negative (T N ) is the percentage of truthful news that was successfully classified utilizing the proposed system, False Positive (F P ) is the percentage of truthful news classified as FNs, and the percentage of FNs items classified as truthful news is represented by False Negative (F N ).Kappa is calculated with the following formula: P o is the model's overall accuracy, and P e is the agreement between the model predic- tions and the actual class values.Precision [74] is expressed as in Eq. ( 29): Recall [75] is expressed as in Eq. ( 30): [76] is obtained as in Eq. ( 31): Specificity [77] is expressed as in Eq. ( 32): Sensitivity [77] is expressed as in Eq. ( 33): (27) The fitness measure calculates the mean fitness results achieved by running the suggested method separately for 30 runs, demonstrating the synergy between minimizing the number of features selected and reducing the error classification rate as Eq.(34).The minimum value presents the best result, which is assessed according to fitness as: where f k • is the optimum fitness result achieved in the k-th run.Features Size measure shows the mean number of selected features by running the method separately 30 runs and is defined as: where |d k • | is the size of features chosen in the optimal solution for the k-th run, and |D| represents the complete size of features in the used benchmark.
Standard Deviation (SD): Corresponding to the measures mentioned above, the final results achieved over the 30 independent runs for each algorithm on every dataset are evaluated and analyzed in terms of stability as: where Y denotes the metric to be measured, Y k * is the value of the metric Y in the k-th run, and µ Y is the average of the metric over the 30 independent runs.

Effect of different components of the proposed IBAVO-AO algorithm for FND
The proposed IBAVO-AO algorithm is compared to the original versions of AVO and AO algorithms to show how this hybridization improves the performance of the IBAVO-AO algorithm.Table 4 displays the results of the proposed IBAVO-AO algorithm and its component algorithms on the utilized ISOT-FNs dataset for FND, in which boldface numbers indicate the best results.

Results analysis of the proposed IBAVO-AO algorithm versus diverse state-of-the-art ML methods and metaheuristic algorithms for FND
In the first part of the analysis of the results, we compared the empirical outcomes of the proposed IBAVO-AO algorithm with some state-of-the-art ML methods on the used ISOT-FNs dataset for FND.For a reliable comparison, the suggested system and selected methods are executed on a framework with identical parameters and tested on the same ISOT-FNs dataset.
Table 5 shows the results of the proposed IBAVO-AO algorithm and other stateof-the-art Ml methods on the utilized ISOT-FNs dataset for FND, where boldface numbers indicate the best results.Table 5 shows that the proposed IBAVO-AO (33) algorithm and ML methods are compared and assessed regarding average accuracy, Kappa, Precision, Recall, and F1 score.According to the results obtained, the overall performance of the proposed IBAVO-AO algorithm was further compared with other ML methods on the extracted attributes.The proposed IBAVO-AO algorithm succeeded in categorizing 92.75% of the news articles.After the suggested system, GNB ranked second with a classifying rate of 81.96% of news articles, but there is a gap of more than 10% between it and the suggested system.As presented in Table 5, the DT and k-NN methods generated the least values in categorizing various news articles.
In the second part of the analysis of the results, we compared the experimental results of the proposed IBAVO-AO algorithm with some known metaheuristic optimization techniques on the used data set for FND.As shown in Table 6, the proposed   6, the proposed IBAVO-AO algorithm succeeded in categorizing 92.7% of the news articles.After the proposed IBAVO-AO algorithm, BAVO ranked second with an average rating of 92.62% of news articles.Also, the stability of the proposed IBAVO-AO algorithm is relatively strong depending on the SD values of the different algorithms.Based on the number of features chosen, it is noted that the proposed IBAVO-AO algorithm comes first by selecting the minimum mean size of attributes on the used ISOT-FNs dataset, followed by the BSSA method.Additionally, the proposed IBAVO-AO algorithm obtained the best exploration capability over other algorithms regarding the mean selected features number, which was confirmed by choosing the least features number on the selected ISOT-FNs dataset.That verifies the capability of the proposed IBAVO-AO algorithm to neglect non-significant search regions and discover the most feasible regions.Therefore, the proposed IBAVO-AO algorithm can minimize the feature search region by identifying the most relevant attributes while preserving the highest classification accuracy.Based on fitness value.It should be noted that the proposed IBAVO-AO algorithm first obtains the minimum mean fitness value on the used ISOT-FNs dataset, followed by the BAVO and BAO methods.Finally, the proposed IBAVO-AO algorithm is based on the remaining evaluation measures.Figure 4 reveals that the proposed approach on the selected attributes generated the highest performance compared to other optimization algorithms.The mean accuracy and F-measure of the proposed IBAVO-AO algorithm are 92.75% and 92.76%, respectively.Following the proposed IBAVO-AO algorithm, BAVO outperforms other known methods with average accuracy and F-measure of 92.62% and 92.62%.The BAO method also achieved nearly comparable outcomes with BAVO, producing accuracy and F-measure of 92.61% and 92.62%.The BSSA algorithm obtained good results by successfully classifying 92.58% of news articles.The BHGSO method obtained the least results and classified 92.11% of news articles.The comparative study results show that the proposed IBAVO-AO algorithm obtained greater F-measure and accuracy than state-of-the-art optimization methods and was reliable in classifying various news articles.

Analysis of convergence
This section reveals an asymptotic investigation of the proposed IBAVO-AO algorithm for handling the FND strategy on the selected dataset to verify its capability in convergence, as shown in Fig. 5.These convergence graphs show the convergence capability of the proposed IBAVO-AO algorithm against their peers, which are all evaluated and executed under identical situations of the number of iterations and population size.Figure 5 shows that the proposed IBAVO-AO algorithm demonstrated fast yet optimal convergence behavior on the selected dataset.Hence, the proposed IBAVO-AO algorithm emphasizes its ability to acquire the optimal solution on time, ensuring an effective balance between exploration and exploitation capabilities.

Comparison results of the proposed IBAVO-AO algorithm against different algorithms from existing studies for FND
Table 7 illustrates the experimental outcomes of comparisons in terms of Accuracy, Kappa, Precision, Recall, and F1-score metrics between the suggested IBAVO-AO algorithm and other algorithms from existing studies, including the WOA-Xgb-Tree [24], AB [12], Wang-CNN [78], Wang-Bi-LSTM [78], Ridor [79,80], and IBk [80,81] for FND    measures.None of the ML models used in the comparison ranked first in any of the performance measures.
In the second part of the analysis of the results, we compared the experimental results of the proposed IBAVO-AO algorithm with some known metaheuristic optimization algorithms on the new unseen datasets for FND, in which boldface values determine the best results.As shown in Table 9, the proposed IBAVO-AO algorithm and wellknown metaheuristic optimization regarding average accuracy, fitness values, number of selected features, Kappa, Precision, Recall, F1-score, Specificity, Sensitivity, ROC_AUC, and MCC.According to the results obtained in Table 9, the proposed IBAVO-AO algorithm outperformed all utilized optimization techniques in all performance measures.None of the optimization techniques used in the comparison ranked first in any of the performance measures.

Conclusion and future work
Recently, FNs have been the most critical issue that harms society and individual users, making FND a great challenge.This study presented a new FNs classification and detection paradigm depending on an effective IBAVO-AO algorithm with the Xgb-Tree classifier.The proposed IBAVO-AO algorithm has preliminary stages: The ISOT-FNs dataset is retrieved first.Then, a pre-processing step is performed to transfer the unstructured data into structured data and analyze and extract the necessary attributes.This step includes extracting attributes from the ISOT-FNs dataset by ignoring useless words, stemming, tokenizing, encoding, and padding data into a sequence of integers using the GLOVE method; the extracted attributes are then filtered utilizing the effective Relief method to discover only suitable ones.Finally, the retrieved features are used to categorize the news items using the proposed IBAVO-AO based on The Xgb-Tree classifier.The suggested system obtained results have been analyzed and compared with state-of-the-art ML classifiers and optimization techniques concerning the accuracy, fitness values, the number of selected features, Kappa, Precision, Recall, F1-score, Specificity, Sensitivity, ROC_AUC, and MCC toward the same ISOT-FNs dataset.Moreover, one focus point was extracting attributes from news articles to assist the FND system in getting higher accuracy and shorter processing time.The results obtained from the proposed IBAVO-AO algorithm showed that the extracted attributes positively affect the performance of the proposed FND system.
Following this, it's important to state the limitations of the study to provide a balanced and realistic view of its scope and applicability: • Dataset Scope and Diversity: The ISOT-FNs dataset, while comprehensive, may not fully encompass the broad spectrum of FNs sources, especially more subtle or complex misinformation strategies.• Single-Modality Focus: The study focused solely on text-based news articles, excluding multimedia elements like images or videos often integral to FNs. • Algorithmic Adaptability: The performance and adaptability of the IBAVO-AO algorithm across various datasets and types of FNs content require further exploration.
In the future, we tend to analyze and investigate these topics: • Incorporating Multi-modal Data: Future research will focus on processing news articles that include images and text, moving beyond the text-only approach to provide a more comprehensive analysis of FNs.• Exploring Diverse Classification Methods: Plans include applying other classification methods like neural networks, k-NN, and Random Forest (RF) to assess further the behavior and efficacy of the IBAVO-AO algorithm in various classification tasks.• Broadening Input Features: We aim to analyze optimization methods with multiple input features, ranging from raw text to handcrafted attributes.This approach could uncover new insights and enhance the system's ability to detect FNs more accurately.

Fig. 1
Fig. 1 Flowchart of the proposed AVO algorithm

Fig. 2
Fig. 2 Flowchart of the proposed AO algorithm

3 Minimum loss reduction gamma = 0 Minimum sum of instance weight min_child_weight = 1 75 DT Maximum depth of a tree max_depth = 5 Number of features max_features = 1 k-NN Euclidean distance metric k = 5 SVM Regularization parameter C = 1 Degree of polynomial kernel degree = 2 RF
-Tree Number of boosting iterations nrounds = 100 Maximum depth of a tree max_depth = Step size shrinkage (learning rate) eta = 0.4 Sub-sample ratio of columns colsample_bytree = 0.8 Sub-sample ratio of training sub_sample = 0.Number of trees in a forest n_estimators = 10 Maximum depth of a tree max_depth = 5 Number of features max_features = 1 MLP Number of neurons in the i th hidden layer hidden_layer_sizes = (1000, 500, 100) Strength of the L2 regularization term alpha = 0.001 Maximum number of iterations max_iter = 1000

Fig. 4
Fig.4 Comparison of the proposed IBAVO-AO algorithm with other optimization algorithms for FND

Fig. 5
Fig. 5 Convergence graph of the proposed IBAVO-AO algorithm and other optimization algorithms on the selected dataset for FND

Table 3
Parameter settings for used meta-heuristic optimization algorithms

Table 4
Results of the proposed IBAVO-AO algorithm and its components for FND

Table 5
Results of the proposed IBAVO-AO algorithm and well-known ML methods for FND

Table 6
Results of the proposed IBAVO-AO algorithm and other optimization algorithms for FND IBAVO-AO algorithm and well-known metaheuristic optimization techniques are compared and evaluated in terms of average accuracy, fitness values, number of selected features, Kappa, Precision, Recall, F1-score, Specificity, Sensitivity, ROC_ AUC, and MCC.Note that boldface values denotes the best results.Regarding classification accuracy values presented in Table

Table 7
Comparison of the proposed IBAVO-AO algorithm and other algorithms from existing studies for FND (Results are reported directly from the studies.)

Table 9
Results of the proposed IBAVO-AO algorithm and other optimization algorithms using other datasets for FND