Skip to main content

A study on improving turnover intention forecasting by solving imbalanced data problems: focusing on SMOTE and generative adversarial networks

Abstract

This study aims to improve the accuracy of forecasting the turnover intention of new college graduates by solving the imbalance data problem. For this purpose, data from the Korea Employment Information Service's Job Mobility Survey (Graduates Occupations Mobility Survey: GOMS) for college graduates were used. This data includes various items such as turnover intention, personal characteristics, and job characteristics of new college graduates, and the class ratio of turnover intention is imbalanced. For solving the imbalance data problem, the synthetic minority over-sampling technique (SMOTE) and generative adversarial networks (GAN) were used to balance class variables to examine the improvement of turnover intention prediction accuracy. After deriving the factors affecting the turnover intention by referring to previous studies, a turnover intention prediction model was constructed, and the model's prediction accuracy was analyzed by reflecting each data. As a result of the analysis, the highest predictive accuracy was found in class balanced data through generative adversarial networks rather than class imbalanced original data and class balanced data through SMOTE. The academic implication of this study is that first, the diversity of data sampling methods was presented by expanding and applying GAN, which are widely used in unstructured data sampling fields such as images and images, to structured data in business administration fields such as this study. Second, two refining processes were performed on data generated using generative adversarial networks to suggest a method for refining only data corresponding to a more minority class. The practical implication of this study is that it suggested a plan to predict the turnover intention of new college graduates early through the establishment of a predictive model using public data and machine learning.

Introduction

As public data disclosure began in many countries, public data have been used in various fields. Recognizing the importance of public data, thus, the South Korean government enacted the ‘Act on Promotion of Provision and Use of Public Data’ in 2013. In the past, public data was used in a limited and exclusive environment for research or policy development purposes. Recently, anyone can easily obtain public data from the government’s portal websites or national institution websites in South Korea. These public data consist of various fields such as disaster safety, education, health and medical care, weather environment, public administration, public finance, national land management, social welfare, employment, food security, and cultural tourism. Therefore, various studies using these public data are being conducted. In particular, as youth unemployment has become a social problem in Korea, studies have been conducted to predict turnover intention and job performance using public data, machine learning, and deep learning to solve this problem [1, 2].

When conducting such a predictive study, the class of variables to be predicted in the data (ex: there is a turnover intention/there is no turnover intention) must be balanced. If the data with unbalanced classes are taught to the predictive model as it is, the model learns only information biased to a large number of classes, so the performance (prediction accuracy) of the model may decrease, and it is difficult to trust the predicted results [3,4,5].

This problem is called the data imbalance problem. Class imbalance data refers to at least one of its classes is usually outnumbered by the other classes. In the field of data mining, detecting events is a prediction problem, or, typically, a data classification problem. Rare events are difficult to detect because of their infrequency and casualness; however, misclassifying rare events can result in heavy costs. For financial fraud detection, invalid transactions may only emerge out of hundreds of thousands of transaction records, but failing to identify a serious fraudulent transaction would cause enormous losses. The scarce occurrences of rare events impair the detection task to imbalanced data classification problem. For example, for a data set where only 1% of the instances belong to the minority class, even if a model classifies all instances as the majority class, it still achieves an overall accuracy of 99%. However, the minority class instances, which we want to accurately classify, are all misclassified by this model though it achieves a very high accuracy.

This class imbalance problems have been reported to occur in a wide variety of real-world domains, such as facial age estimation [6], detecting oil spills from satellite images [7], anomaly detection [8], identifying fraudulent credit card transactions [9], software defect prediction [10], and image annotation [11]. Beurz and Van den Poel [3] suggested sampling techniques and algorithms to solve the problem, noting that class imbalance in learning data leads to negative results in modeling for customer departure prediction. Haixiang et al. [4] reviewed 527 articles related to unbalanced data published over the past decade. In particular, they found that although the imbalance data problem may occur in the field of business administration, there are not many papers that solve the imbalance problem in this area. Seliya et al. [5] mentioned imbalanced data as a unique problem that can occur in big data, and investigated how the biased learning of imbalanced data leads to negative results in binary classification problems by investigating related studies for 10 years.

As Haxiang et al. [4] mentioned, predictive modeling using big data is also being carried out in the field of business administration. Applying this technique to the field of personnel organization can provide important insights in predicting the turnover intention of new employees, which has recently become a problem. Recently, the problem of youth employment, especially the deterioration of employment for college graduates, has become a social issue. However, companies are also complaining of a shortage of human resources, and one of the causes is the frequent turnover of new employees with college graduates [12]. According to the Ministry of Trade, Industry and Energy [13], the early retirement rate of experienced people is 13%, while the early retirement rate of new employees is 66%. According to a survey of economic activities by Statistics Korea [14], if young people aged 15–29 quit their first job, the average service period is 1.9 months. Most young people want to get a job, but those who succeed in getting a job are showing a paradoxical form of considering changing jobs or quitting. However, it is not easy for companies to measure the turnover intention of new employees. Because the turnover intention is a sensitive matter to individuals, and it is difficult to disclose it easily due to concerns about the personnel disadvantages that may arise when it is disclosed to the outside [15]. Also, even if it disclosed, it is not easy to determine whether the content is true. Companies are facing difficulties in selecting and managing new employees with turnover intentions.

Therefore, this study proposes a method to predict the turnover intention of new employees. If a predictive model is built using public data and machine learning algorithms that measure the turnover intention of new college graduates, and the independent variables used in model construction are measured on new employees and put into the model, their turnover intention can be predicted. This approach not only provides practical implications for promptly identifying new employees with turnover intentions in situations where it is challenging to easily obtain sensitive information such as turnover intention, but also supplements the limitations of traditional econometric models that solely focus on causal analysis. However, even in the process of building such a prediction model, class imbalance problems may still arise. For example, if there are 10 learning data, consisting of a ratio of 8 employees who are willing to change jobs and 2 employees who are not willing to change jobs, the model is likely to predict employees who are not willing to change jobs at 80%. This may result in fairness issues in the selection process for new employees who are willing to turnover, and could potentially have negative impacts on organizational management. Thus, it is crucial to address the class imbalance problem in the training data when building a predictive model. Therefore, in this study, a solution to the problem of imbalanced data was presented through oversampling using SMOTE and generative adversarial networks. In order to check the effect of solving the imbalanced data problem, the predictive model was analyzed by the original data with unbalanced classes and then modified data via two methodologies: SMOTE and generative adversarial network are also applied. Then, the results were compared. The contributions of this study are as follows.

  • the diversity of data sampling methods was presented by expanding and applying generative adversarial networks, which are widely used in unstructured data sampling fields such as images and images, to structured data in business administration fields such as this study.

  • two refining processes were performed on data generated using generative adversarial networks to suggest a method for refining only data corresponding to a more minority class.

  • this study suggested a plan to predict the turnover intention of new college graduates early through the establishment of a predictive model using public data and machine learning.

Background and literature review

Imbalanced data

Imbalanced data refers to a case where the number of data belonging to one class is significantly larger or smaller than the number of data belonging to another class. At this time, a class with a larger number of data is called a majority class, and a class with relatively small data is called a minority class [16]. There is an imbalanced ratio (IR) to indicate the degree of data imbalance, and it is defined as follows.

$$IR=\frac{{n}_{maj}}{{n}_{min}}$$
(1)

Here, nmaj and nmin refer to the number of data belonging to the majority class and the minority class. If IR = 3, it means that the data belonging to the majority class is three times more than the data belonging to the minority class. Using these imbalanced data for prediction purposes can cause many problems [17]. For example, when learning a prediction model using machine learning, if the class of the target variable is imbalanced in the training data (e.g., when the class of the target variable is 0 or 1, the model is biased toward the class that occupies a large proportion [18]. As a result, this negatively affects the performance (prediction accuracy) of the model [3, 4]. This is called the class imbalance problem, and it is a problem that must be dealt with caution in prediction research, especially in finance, medical care, and manufacturing industry [19]. For example, in marketing, when making a model to determine whether or not a customer has churned, training the model through measured customer data, and then reflecting new customer data to predict the likelihood of their churn [20]. If most of the training data consist of customers who do not churn, the model that learned this has a problem of classifying new customers as customers who do not leave most of the time.

In general, a minority class is a subject of major interest in classification problems [21], However, when a class imbalance problem occurs, it is difficult to obtain satisfactory classification performance, and in some cases, user damage may occur due to incorrect classification of minority class data [22]. For example, if a patient with a rare disease is incorrectly classified as a normal person in the medical field, the opportunity to treat the patient is lost. In addition, additional damage may occur if payment history with a stolen credit card is incorrectly classified as normal in fraud detection in the financial field [23]. Since it is very important to accurately classify a minority class in this way, it is necessary to solve the problem when the class of data used is imbalanced. Data sampling is a method to solve the problem of imbalanced data [24]. Data sampling is divided into under-sampling, which deletes data by matching the majority class to the size of the minority class, and over-sampling, which generates data by matching the minority class to the size of the majority class [24]. Undersampling can save time and cost of additional data collection but has the disadvantage of losing helpful information of the data [25]. On the other hand, oversampling may increase the time required to train the predictive model due to additional data generation, but it is possible to avoid the loss of important information in the collected data. Figure 1 shows the process of undersampling and oversampling.

Fig. 1
figure 1

Data sampling process [16]

Since one of various sampling techniques cannot be said to be better, and each technique has its pros and cons, it is necessary to utilize a technique suitable for a given data and task. But undersampling removes data, such methods risk the loss of important concepts. We focused on oversampling, not undersampling, to avoid loss of important information. Among the oversampling techniques, there is also a method of increasing the number of data through simple randomization [26]. However, this method can also lead to overfitting problems because it simply copies data. On the other hand, SMOTE generates data based on algorithms, so the likelihood of overfitting occurrence is less than that of simple random methods. As a consequence, in the field of imbalance data classification problem, a better performance can be easily achieved [27]. In addition, prior studies that solve the imbalance problem through oversampling have used SMOTE in their research, emphasizing that it is the most popular model among the proposed methods [27,28,29,30,31,32,33,34,35]. In particular, SMOTE is used as an imbalance class processing method for data measured on the Likert scale as used in this study [36], and there is evidence that it is also used as an oversampling method for the development of a turnover intention prediction model [37,38,39].

Synthetic minority oversampling technique (SMOTE)

SMOTE is a technique for generating data through bootstrapping and the K-nearest neighbor methods. For specific data belonging to a minority class, K nearest neighbors of the same minority class are found, and new data are generated between them by creating a linear connection structure with the neighbors [40]. This is shown in the following Fig. 2.

Fig. 2
figure 2

SMOTE [40]

Studies have been conducted to solve the class imbalance problem through such SMOTE. Wang et al. [28] applied SMOTE to address the problem of unbalanced data in healthcare. He et al. [29] proposed the use of SMOTE to solve the imbalance problem in various research fields. Chen [41] used several re-sampling techniques for finding the maximum accuracy of classification from fully labeled imbalanced training data set. SMOTE, Oversampling by duplicating minority examples, random under sampling, is mainly used to create new training data set. Rahman and Davis [30] tried to address class imbalance issue in medical datasets. They used undersampling techniques as well as oversampling techniques like SMOTE to balance the classes. Jishan et al. [31] proposed a method of preprocessing data to improve the accuracy of students' final grade prediction models using SMOTE. Douzas et al. [32] proposed an unbalanced data solution using SMOTE for the unbalanced dataset provided by UCI. Feng et al. [33, 34] proposed an SMOTE-based oversampling method to alleviate the class imbalance problem in software defect prediction. Nunez and Gatica [35] utilized SMOTE to solve the imbalance data problem in the Peruvian public investment prediction model.

Like this, SMOTE has been used in many fields to solve imbalanced data problems. However, when SMOTE generates data belonging to a minority class, it does not consider the location of data belonging to the adjacent majority class. Therefore, the problem of overlapping positions occurs [42]. In addition, since data is generated by relying only on the relationship between a few classes, overfitting may occur and prediction performance may deteriorate [43]. Therefore, it is necessary to generate new data that does not overlap with the existing data while considering the overall distribution of the data when generating data was derived [44]. To overcome these issues, we adopted the generative adversarial network (GAN) concept. GAN can solve overfitting and data superposition problems because it learns the actual data distribution of minor class and then generates similar data [45]. Through this, GAN can overcome the limitations of existing oversampling techniques.

Generative adversarial network (GAN)

GAN is an unsupervised learning-based neural network that learns and optimizes by competing with a generator that generates data and a discriminator that compares the generated data with real data [46]. Since the two neural networks compete with each other, the quality of the generated data increases as learning takes place, and when the generator is optimized, the generated data becomes very similar to the quality of real data. As shown in the following Fig. 3, if a random type of noise (Z) is put into the generator, the generator generates fake data based on it. When this is put into the discriminator, the discriminator compares it with the existing actual data to determine whether the data generated by the generator is real or not. The result is expressed as a loss value, and when this value is sent to the generator, the generator proceeds with learning to generate fake data similar to real data based on this. After that, fake data is generated and sent back to the discriminator. This process proceeds until the discriminator does not clearly distinguish between real data and fake data, and when this state is reached, the generator is said to be optimized [45].

Fig. 3
figure 3

Generative adversarial networks [46]

The goal of the generator is to maximize the probability (D(G(z))) that the discriminator will distinguish fake data into real data by 1. The goal of the discriminator is to maximize the probability (D(x)) of determining that the generated fake data is fake by comparing it with the actual data. The formula is as follows.

$$\begin{array}{c}Min\\ G\end{array}\begin{array}{c}Max\\ D\end{array}\begin{array}{c}V(D, G) = {E}_{{x} \sim Pdata({x})}[logD(\text{x})]+{E}_{{z}\sim P\text{z}({Z})}[log(1-D(G({z})))]\\ \end{array}$$
(2)

GAN has been mainly used in the field of unstructured data such as images and videos [47]. However, it has also been used in the field of database sized structured data. In particular, studies applied to oversampling for solving the class imbalance problem have taken placed. Kim [48] presented a solution to the problem of data imbalance for predicting defaulted companies through generative adversarial networks based oversampling. Kim et al. [49] also suggested a solution to the problem of data imbalance for predicting credit card fraud through generative adversarial networks based oversampling. Mao et al. [50] presented a solution to the problem of data imbalance for predicting defective products in the manufacturing process through generative adversarial networks based oversampling. Table 1 shows the summary of these studies.

Table 1 Research on generative adversarial networks

Research methodology

Introduction to the dataset

This study is to solve the class imbalance data problem and propose a method to improve the accuracy of the prediction model. For these purposes, data from the Graduate Occupations Mobility Survey (GOMS) of the Korea Employment Information Service was used. This data includes the turnover intention composed of class variables and personal characteristics, job characteristics, characteristics of the university, job search experience, and so on. The data consisted of 18,163 samples and 1270 variables. The data used in this study is freely available to anyone (Data download: https://github.com/jrpark16/GOMS.git). In order to predict turnover intention, 12,202 samples of respondents who worked in the past 4 weeks were extracted. The employee turnover intention, which is the dependent variable, was obtained from these samples as a discrete class. To the question of turnover intention, 2862 (23.46%) of the respondents answered 'Yes,' and 9340 (76.54%) answered 'No.' Fig. 4 shows that these classes are imbalanced data.

Fig. 4
figure 4

Turnover intention rate

Variables

Thirteen explanatory variables were derived from the viewpoint of job choice motivation, needs satisfaction, and person-job fit based on previous studies. The following Table 2 shows the summarized variables.

Table 2 Variables

Oversampling

In this study, oversampling was performed through SMOTE and generative adversarial networks to match the imbalanced class ratio of the dependent variable in the original data. First, the original data was divided into 70% training and 30% test sets to avoid overfitting. Since then, the training set has divided data in response to "I have turnover intention" and data in response to "I do not have turnover intention." After that, the data in response to "I have turnover intention" was used as data for oversampling. Afterward, the class ratio of the dependent variable was matched by oversampling the previously classified data through the SMOTE algorithm provided by the Imbalanced-learn library. Next, GANs were built and trained on the previously classified data, and data oversampling was performed based on the optimized generator to match the class ratio of the dependent variable. The structure of the constructed GANs are as follows.

First, the generator receives a noise vector Z as input. Since the input Z is irrelevant to the output data, a dimension sufficient to contain the properties of the data may be arbitrarily set. This study was organized in 50 dimensions so that the data distribution in the latent space could be properly trained. The generator layer comprises Z(50) → 32 → 64 → 128 → 13 (number of explanatory variables). The discriminator layer is composed of the number of generator outputs (13) → 128 → 64 → 32 → 16 → 1 (sigmoid). This is shown in the following Fig. 5.

Fig. 5
figure 5

Generator and discriminator layer

The batch size of the learning process was 50 and the epoch was performed 3,000 times. After learning 2000 times, the loss values of generator and discriminator did not change significantly around 0.7, so only 3000 times were performed and learning process was terminated. Then, data was generated through the learned generator. The learning process is shown in Fig. 6.

Fig. 6
figure 6

Learning process

The generated data went through two purification processes. First, a prediction model for the turnover intention was trained through the original data, then the generated data was put in and analyzed through XGBoost, and only the data belonging to ‘I have turnover intention’ was primarily refined. After analyzing the refined data through logistic regression, only the data belonging to ‘I have turnover intention’ was secondarily refined. By including the ‘I have turnover intention’ data that had undergone two refinements in the imbalanced training set, the class ratio of the dependent variable in the training set was made the same. Through these processes, class imbalanced original data, class balanced data through SMOTE, and class balanced data through generative adversarial networks were prepared.

Turnover intention prediction analysis

In order to predict turnover intention, a prediction model was trained using the supervised learning method during machine learning and then the prediction accuracy was analyzed. For this, the following method was performed. First, a prediction model for turnover intention was constructed in which all the variables presented in Table 2 were set as explanatory variables, and turnover intention was set as the dependent variable. Second, the three previously prepared data were divided into 70% training set and 30% test set, respectively, and after learning the turnover intention prediction model with the training set, the model's prediction accuracy was analyzed with the test set. Third, the prediction model was analyzed using logistic regression (LR), K-nearest neighbor (KNN), and XGBoost (eXtreme Gradient Boosting: XGB). Accuracy, Precision, Recall, and F1-score were used as prediction performance evaluation indicators. In addition, cross-validation was performed four times to prevent the error of overfitting the prediction model only to a specific data set and to generalize the analysis results.

Meanwhile, Cross-validation was performed four times to prevent the model from overfitting only a specific data set and to generalize the analysis results. On the other hand, the widely used Accuracy, Precision, Recall, and F1-score are used as the performance evaluation indicators of the predictive model. The meaning and formula of each indicator are as follows.

Accuracy: Accuracy is the most intuitive performance measure and is simply a ratio of correctly predicted observations to the total observations.

$$\left(\mathrm{Accuracy}\right)= \frac{\mathrm{TP}+\mathrm{TN}}{\mathrm{TP}+\mathrm{FN}+\mathrm{FP}+\mathrm{TN}}$$
(3)

Precision: Precision is the ratio of correctly predicted positive observations to the total predicted positive observations.

$$\left(\mathrm{Precision}\right)= \frac{\mathrm{TP}}{\mathrm{TP}+\mathrm{FP}}$$
(4)

Recall: Recall is the ratio of correctly predicted positive observations to all observations in the actual “yes” class.

$$\left(\mathrm{Recall}\right)=\frac{\mathrm{TP}}{\mathrm{TP}+\mathrm{FN}}$$
(5)

F1-score: The F1-score is the weighted average of precision and recall.

$$\left(F1{\text{-}}score\right)=2 \times \frac{1}{\frac{1}{Precision} + \frac{1}{Recall}} =2 \times \frac{Precision \times Recall }{Precision+Recall}$$
(6)

Results and discussion

First, looking at the average of the results of 4 cross-validation after analysis using class imbalanced original data, LR showed accuracy of Accuracy = 0.783, Precision = 0.798, Recall = 0.956, and F1-score = 0.870. KNN showed accuracy of Accuracy = 0.761, Precision = 0.798, Recall = 0.915, and F1-score = 0.853. XGB showed accuracy of Accuracy = 0.785, Precision = 0.806, Recall = 0.942, and F1-score = 0.869.

Next, looking at the average of the results of 4 cross-validation after analysis using class balanced data using SMOTE, LR showed accuracy of Accuracy = 0.782, Precision = 0.797, Recall = 0.955, F1-score = 0.869. KNN showed accuracy of Accuracy = 0.765, Precision = 0.802, Recall = 0.914, and F1-score = 0.854. XGB showed accuracy of Accuracy = 0.784, Precision = 0.806, Recall = 0.941, and F1-score = 0.891. As a result of the analysis, it was found that there was no significant difference in prediction accuracy between the class imbalanced original data and the data with class balanced using SMOTE.

Next, looking at the average of the results of 4 cross-validation after analysis using class balanced data using GAN, LR showed accuracy of Accuracy = 0.799, Precision = 0.824, Recall = 0.937, F1-score = 0.877. KNN showed accuracy of Accuracy = 0.794, Precision = 0.820, Recall = 0.929, and F1-score = 0.871. XGB showed accuracy of Accuracy = 0.823, Precision = 0.838, Recall = 0.951, and F1-score = 0.891.

Comparing the average accuracy with the class imbalanced original data, in the case of LR, the balanced data with GAN showed higher prediction accuracy of Accuracy = 1.6%, Precision = 2.6% and F1-score = 0.7% than the original data. However, in Recall, the original data showed high prediction accuracy by 1.9%. In the case of KNN, the balanced data with GAN showed higher prediction accuracy of Accuracy = 3.3%, Precision = 2.2%, Recall = 1.4%, and F1-score = 1.8% than the original data. In the case of XGB, the balanced data with GAN showed higher prediction accuracy of Accuracy = 3.8%, Precision = 3.2%, Recall = 0.9%, and F1-score = 2.2% than the original data. As a result of the analysis, the highest prediction accuracy was shown when class balanced data using GAN was used, and there was no significant difference in prediction accuracy between the original data and class-balanced data using SMOTE. To summarize this, Table 3 is as follows.

Table 3 Prediction analysis result

According to the results of this study, it was found that there was no significant difference in prediction accuracy when the original data in which the class ratio of the dependent variable was unbalanced and the data in which the class ratio was balanced through oversampling using SMOTE were used for prediction model analysis. However, this analysis result does not consider the data values belonging to multiple classes when oversampling through SMOTE, so the generated data overlap [42] and may overfit [43] consistent with the result. Therefore, when oversampling data using SMOTE, it is necessary to consider the overall distribution of prime class data, such as increasing the number of oversampling times or increasing the number of K that can be considered. In addition, after data oversampling, it is necessary to determine whether there are overlapping values through comparison with data belonging to multiple classes.

On the other hand, it was found that the prediction accuracy was higher than that of the original data when data balanced with class ratios were used for prediction model analysis through oversampling using GANs. These analysis results are consistent with previous studies [49, 51, 53] that prediction accuracy is higher when data with imbalanced class ratio is used for prediction model learning. When the model is trained on data with disproportionate class proportions, the accuracy is very high. However, the recall of classes with a small number of data rapidly decreases. One of the practical implications of this study is to suggest a way to predict employee turnover intention based on a pre-trained turnover intention prediction model. Therefore, it can be seen that the recall of the model is a very important study. Previous studies dealing with the class imbalance problem claim that class imbalance through oversampling can consequently increase the model recall [54,55,56]. Khoda et al. [54] compared oversampling and undersampling, and confirmed that solving the class imbalance problem through oversampling resulted in a higher F1 score. Prasetiyo et al. [55] found that the recall value was highest when the classes were balanced 50:50 through oversampling. Jo and Kim [56] suggested that in binary classification problems in various industries such as business, medical, and marketing, models learned from data with class balance through oversampling show the highest recall. In addition, it is believed that this is because similar but non-overlapping data were generated while considering the overall distribution of minority data during data oversampling. Therefore, when analyzing, it is necessary to check the class ratio of the dependent variable in the data if the dependent variable consists of a discrete class. If the class ratio is imbalanced, it is necessary to balance it through data oversampling methods. In addition, if the analysis is performed by applying a method of generating non-overlapping data while considering the overall distribution of minority data, higher prediction accuracy can be derived.

Conclusion

This study focused on solving unbalanced data problems to derive higher predictive results in situations where data disclosure and sharing are active and artificial intelligence technology to analyze them is being developed and used for prediction in various fields. In addition, it was suggested that the performance of the predictive model can be improved when solving class imbalance using GANs. The academic implications of this study are that first, the diversity of data sampling methods was presented by expanding and applying GANs, which are widely used in unstructured data sampling fields such as images and videos, to structured data in business administration fields. Second, we proposed to perform two refinement processes through logistic regression analysis and XGBoost as a way to check whether the sampled data contains a small number of data properties well. The practical implication of this study is that it suggested a plan to predict the turnover intention of new college graduates early through the establishment of a predictive model using public data and machine learning. The main purpose of this study was to propose a method to improve the accuracy of the prediction model by resolving data imbalance through SMOTE, a data oversampling method, and GANs. There are limitations to its use. There are various parameters that can be applied to each classifier. Therefore, future research can be performed with an optimization algorithm such as grid search to find and apply optimal parameters, and in this way, higher prediction accuracy is expected.

Availability of data and materials

https://github.com/jrpark16/GOMS.git.

Abbreviations

GAN:

Generative adversarial networks

SMOTE:

Synthetic minority oversampling technique

LR:

Logistic regression

KNN:

K-nearest neighbor

XGB:

EXtreme Gradient Boosting

References

  1. Lee E-J, Cho H-S, Song Y-S. An exploratory study on determinants predicting university graduate newcomers’ early turn over. J Corporate Educ Talent Res. 2020;22(1):163–93.

    Article  Google Scholar 

  2. Choi J-W, Shin D-W, Lee H-J. Turnover rate prediction among IT firms according to job satisfaction and dissatisfaction factors: using topic modeling and machine learning. J Korean Data Inf Sci Soc. 2021;32(5):1035–47.

    Google Scholar 

  3. Burez J, Van den Poel D. Handling class imbalance in customer churn prediction. Expert Syst Appl. 2009;36:4626–36.

    Article  Google Scholar 

  4. Haixiang G, Yijing L, Shang J, Mingyun G, Yuanyue H, Bing G. Learning from class-imbalanced data: review of methods and applications. Expert Syst Appl. 2017;73:220–39.

    Article  Google Scholar 

  5. Seliya N, Zadeh A-A, Khoshgoftaar T-M. A literature review on one-class classification and its potential applications in big data. J Big Data. 2021;8:122.

    Article  Google Scholar 

  6. Chao WL, Liu JZ, Ding JJ. Facial age estimation based on label-sensitive learning and age-oriented regression. Pattern Recogn. 2013;46(3):628–41.

    Article  Google Scholar 

  7. Kubat M, Holte RC, Matwin S. Machine learning for the detection of oil spills in satellite radar images. Mach Learn. 1998;30:195–215.

    Article  Google Scholar 

  8. Khreich W, Granger E, Miri A, Sabourin R. Adaptive ROC-based ensembles of HMMs applied to anomaly detection. Pattern Recogn. 2012;45(1):208–30.

    Article  MATH  Google Scholar 

  9. Fawcett T, Provost F. Adaptive fraud detection. Data Min Knowl Disc. 1997;1(3):291–316.

    Article  Google Scholar 

  10. Pelayo L, Dick S. Applying novel resampling strategies to software defect prediction. In: NAFIPS 2007–2007 annual meeting of the North American fuzzy information processing society. New York: IEEE; 2007. p. 69–72.

  11. Zhang D, Islam MM, Lu G. A review on automatic image annotation techniques. Pattern Recogn. 2012;45(1):346–62.

    Article  Google Scholar 

  12. Chung DB. Major factors affecting turnover intention of college graduates: comparison and analysis according to regular workers. Q J Labor Policy. 2019

  13. Ministry of Trade, Industry and Energy. A survey on the supply and demand trend of industrial technology personnel in industrial technology. 2017.

  14. Statistics Korea. The results of an additional survey of young people in the May 2018 economically active population survey. 2018.

  15. Mobley WH. Some unanswered questions in turnover and withdrawal research. Acad Manag Rev. 1982;7(1):111–6.

    Article  Google Scholar 

  16. Sun Y, Kamel M-S, Wong A-K-C, Wang Y. Cost-sensitive boosting for classification of imbalanced data. Pattern Recogn. 2007;40(12):3358–78.

    Article  MATH  Google Scholar 

  17. Johnson J-M, Khoshgoftaar T-M. Survey on deep learning with class imbalance. J Big Data. 2019;6:27.

    Article  Google Scholar 

  18. O’Brien R, Ishwaran H. A random forests quantile classifier for class imbalanced data. Pattern Recogn. 2019;90:232–49.

    Article  Google Scholar 

  19. Tanha J, Abdi Y, Samadi N, Razzaghi N, Asadpour M. Boosting methods for multi-class imbalanced data classification: an experimental review. J Big Data. 2020;7:70.

    Article  Google Scholar 

  20. Ahmad A-K, Jafar A, Aljoumaa K. Customer churn prediction in telecom using machine learning in big data platform. J Big Data. 2019;6:28.

    Article  Google Scholar 

  21. Leevy J-L, Khoshgoftaar T-M, Bauder R-A, Seliya N. A survey on addressing high-class imbalance in big data. J Big Data. 2018;5:42.

    Article  Google Scholar 

  22. Hasanin T, Khoshgoftaar T-M, Leevy J-L, Bauder R-A. Severely imbalanced Big Data challenges: investigating data sampling approaches. J Big Data. 2019;6:107.

    Article  Google Scholar 

  23. Benchaji I, Douzi S, Ouahidi B-E, Jaafari J. Enhanced credit card fraud detection based on attention mechanism and LSTM deep model. J Big Data. 2021;8:151.

    Article  Google Scholar 

  24. Hulse J-V, Khoshgoftaar T-M, Napolitano A-N. Experimental perspectives on learning from imbalanced data. In: Proceedings of the ACM international conference on machine learning. 2007. p. 935–42.

  25. Kim H-Y, Lee W-J. On sampling algorithms for imbalanced binary data: performance comparison and some caveats. Korean J Appl Stat. 2017;30(5):681–90.

    Article  MathSciNet  Google Scholar 

  26. Kaggle. https://www.kaggle.com/code/rafjaa/resampling-strategies-for-imbalanced-datasets/notebook.

  27. Chawla N-V, Bowyer K-W, Hall L-O, Kegelmeyer W-P. SMOTE: synthetic minority over-sampling technique. J Artif Intell Res. 2002;16:321–57.

    Article  MATH  Google Scholar 

  28. Wang J, Xu M, Wang H, Zhang J. Classification of imbalanced data by using the SMOTE algorithm and locally linear embedding. In: 2006 8th international conference on signal processing, vol. 3. New York: IEEE; 2006.

  29. He H, Bai Y, Garcia EA, Li S. ADASYN: adaptive synthetic sampling approach for imbalanced learning. In: 2008 IEEE international joint conference on neural networks (IEEE world congress on computational intelligence). New York: IEEE; 2008. p. 1322–8.

  30. Rahman MM, Davis DN. Addressing the class imbalance problem in medical datasets. Int J Mach Learn Comput. 2013;3(2):224.

    Article  Google Scholar 

  31. Jishan ST, Rashu RI, Haque N, Rahman RM. Improving accuracy of students’ final grade prediction model using optimal equal width binning and synthetic minority over-sampling technique. Decis Anal. 2015;2:1–25.

    Article  Google Scholar 

  32. Douzas G, Bacao F, Last F. Improving imbalanced learning through a heuristic oversampling method based on k-means and SMOTE. Inf Sci. 2018;465:1–20.

    Article  Google Scholar 

  33. Feng S, Keung J, Yu X, Xiao Y, Bennin KE, Kabir MA, Zhang M. COSTE: complexity-based OverSampling TEchnique to alleviate the class imbalance problem in software defect prediction. Inf Softw Technol. 2021;129:106432.

    Article  Google Scholar 

  34. Feng S, Keung J, Yu X, Xiao Y, Zhang M. Investigation on the stability of SMOTE-based oversampling techniques in software defect prediction. Inf Softw Technol. 2021;139:106662.

    Article  Google Scholar 

  35. Nunez NA, Gatica G. Applying profit-driven metrics in predictive models: a case study of the optimization of public funds in Peru. J Syst Manag Sci. 2022;12(2):52–65.

    Google Scholar 

  36. Aggarwal S, Saluja S, Gambhir V, Gupta S, Satia SPS. Predicting likelihood of psychological disorders in PlayerUnknown’s Battlegrounds (PUBG) players from Asian countries using supervised machine learning. Addict Behav. 2020;101:106132.

    Article  Google Scholar 

  37. de Oliveira JM, Zylka MP, Gloor PA, Joshi T. Mirror, mirror on the wall, who is leaving of them all: predictions for employee turnover with gated recurrent neural networks. Collaborative innovation networks: latest insights from social innovation, education, and emerging technologies research. 2019. p. 43–59.

  38. Tao Z, Wu C, Zhao S. Research on the prediction of employee turnover behavior and its interpretability. In: Proceedings of the 2021 5th international conference on electronic information technology and computer engineering. 2021. p. 760–7.

  39. Şahinbaş K. Employee promotion prediction by using machine learning algorithms for imbalanced dataset. In: 2022 2nd international conference on computing and machine intelligence (ICMI). New York: IEEE; 2022. p. 1–5.

  40. Hu F, Li H. A novel boundary oversampling algorithm based on neighborhood rough set model: NRSBoundary-SMOTE. Math Probl Eng. 2013;2013:1–10.

    Article  MathSciNet  MATH  Google Scholar 

  41. Chen Y. Learning classifiers from imbalanced, only positive and unlabeled data sets. Ames: Department of Computer Science, Iowa State University; 2009.

    Google Scholar 

  42. Santoso B, Wijayanto H, Notodiputro K-A, Sartono B. Synthetic over sampling methods for handling class imbalanced problems: a review. In IOP Conference series: earth and environmental science, vol. 58. 2017. p. 1–8.

  43. Krawczyk B. Learning from imbalanced data: open challenges and future directions. Prog Artif Intell. 2016;5(4):221–32.

    Article  Google Scholar 

  44. Bagui S, Li K. Resampling imbalanced data for network intrusion detection datasets. J Big Data. 2021;8:6.

    Article  Google Scholar 

  45. Goodfellow I-J, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y. Generative adversarial networks, In Proceedings of the neural information processing systems. 2014. p. 2672–80.

  46. Kalin J. Generative adversarial networks cookbook, Packt. 2018.

  47. Sampath V, Maurtua I, Aguilar Martín J-J, Gutierrez A. A survey on generative adversarial networks for imbalance problems in computer vision tasks. J Big Data. 2021;8:27.

    Article  Google Scholar 

  48. Kim YL. GAN-based oversampling technique for imbalanced bankruptcy data processing, Ewha Womans University Master’s thesis. 2020.

  49. Kim Y-W, You Y-L, Choi H-Y. Fraud detection system model using generative adversarial networks and deep learning. Inf Syst Rev. 2020;22(1):59–72.

    Google Scholar 

  50. Mao Q, Lee H-Y, Tseng H-Y, Ma-S, Yang M-H. Mode seeking generative adversarial networks for diverse image synthesis. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (CVPR). 2019. p. 1429–37.

  51. Park J-S, Ahn G-S, Heo S. Oversampling based on k-NN and GAN for effective classification of class imbalance dataset. J Korean Inst Ind Eng. 2020;46(4):365–71.

    Google Scholar 

  52. Engelmann J, Essmann S. Conditional Wasserstein GAN-based oversampling of tabular data for imbalanced learning. Expert Syst Appl. 2021;174(15):114582.

    Article  Google Scholar 

  53. Son M-J, Jung S-W, Hwang I-J. A deep learning based over-sampling scheme for imbalanced data classification. KIPS Trans Softw Data Eng. 2019;8(7):311–6.

    Google Scholar 

  54. Khoda M-E, Kamruzzaman J, Gondal I, Imam T, Rahman A. Malware detection in edge devices with fuzzy oversampling and dynamic class weighting. Appl Soft Comput. 2021;112:107783.

    Article  Google Scholar 

  55. Prasetiyo B, Muslim M-A, Baroroh N. Evaluation performance recall and F2 score of credit card fraud detection unbalanced dataset using SMOTE oversampling technique. In: Journal of physics: conference series, vol. 1918, no. 4. Bristol: IOP Publishing; 2021. p. 042002.

  56. Jo W, Kim D. OBGAN: minority oversampling near borderline with generative adversarial networks. Expert Syst Appl. 2022;197:116694.

    Article  Google Scholar 

Download references

Acknowledgements

Not applicable.

Funding

Not applicable.

Author information

Authors and Affiliations

Authors

Contributions

JRP: designed the idea and experiments, and he wrote the manuscript and implemented the code. SDK: provided ideas during the analysis process. JSP has finalized the structure and content of the manuscript. All authors read and approved the final manuscript.

Authors information

JRP is a post-doctoral researcher at Electronics and Telecommunications Research Institute (ETRI), Technology Policy Research Division. He acquired a doctorate in Management Information Systems from Chungbuk National University. His main areas of interest are Machine Learning and Deep Learning, Data Analysis, and Business Prediction.

SDK is currently a professor in MIS program at Chungbuk National University. He acquired a doctorate in Management Information systems from Seoul National University. The main areas of interest are SCM-based Smart Factory, Machine Learning, and Deep Learning-based Business Predictions.

SPJ is currently an associate professor at the Division of Science and Technology of BNU-HKBU United International College. His current interests are Recommender Systems, Business Intelligence Systems, and Data Analysis.

Corresponding author

Correspondence to Seon-Phil Jeong.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

The publisher has the author's permission to publish this paper.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Park, J., Kwon, S. & Jeong, SP. A study on improving turnover intention forecasting by solving imbalanced data problems: focusing on SMOTE and generative adversarial networks. J Big Data 10, 36 (2023). https://doi.org/10.1186/s40537-023-00715-6

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s40537-023-00715-6

Keywords