1. INTRODUCTION
The convenience of credit cards is common in modern day community. Credit card utilization has expanded among the clients since credit card installment is key one and it is helpful to pay the amount. It is utilized either online or conventional shopping. Due to the expansion and fast advancement in the fields such as Ecommerce, the utilization of credit card is also expanded radically [1]. As the use of credit card is development, the credit card fraud is additionally increments. The fraud is characterized as a restricted movement by a client for whom the record was not anticipated [2]. The clients who are utilizing the credit card not having the associations with the cardholder and has no goal of making the repayments for the obtain they done. At present, commercial fraud is turning into a serious issue, and successful identification of credit card is a troublesome effort for the experts [3].
Identifying credit card fraud is a tough effort when applying traditional methods; therefore, the growth of the credit card fraud discovery model has matured off significance, either in the educational or trades the society recently.
Credit card fraud detection belongs to the classification and identification problem with a large number of nonlinear situations, which cause it significant to consider nonlinear integrated ways to explaining the problem [4].
Artificial neural network (ANN) is a mathematical description of the network of neurons in the mind and share relationships functionalities, such as accepting inputs, processing it, and then produces output [5]. It follows a combined graph of nodes, which are twisted by the weighted links related to the biological neurons. There are different models ANN, for example, feedforward neural network (FFNN), multiplelayered perceptron, and Kohonen network. Adaptive resonance network and the initial two nets work as a classifier, i.e. these can learn from patterns, and the knowledge can be immediately supervised. Although the other nets learn from attention and later update the network weights, through serve unsupervised learning system seen in a case of clustering. In this paper, a FFNN has been improved for classification intention. FFNN allows the information to pass from the input to output layer in a feedforward path through the hidden layer(s) [6]. All FFNNs, as stated, possibly trained in a supervised way so that it can learn the feature pattern accessible within the data. To achieve the wanted accuracy in class prediction, fit training is compulsory. While training, the purpose is to catch the network learning feature as the best, which is mirrored by reducing the squared error (i.e., the squared variation between the calculated and the wanted output).
There are various algorithms to optimize such learning method. Backpropagation (BP) is one of the standard traditional ANN training algorithms for supervising learning. The weights are adjusted and updated with a statement delta rule to minimize the prediction error during iterations. The weight improvement methodology covers BP the errors from output layer into hidden layer, so obtaining the optimal set of weights [7].
Simulated annealing (SA) is a probabilistic metaalgorithm for global optimization [8]. It is parallel to the physical method where a solid is casually begin cooled till it is construction is in a cold state, which occurs at a minimum energy form [9]. Similarly, BP algorithm, in SA, the weight has to go into some configuration on the rule till it leads the global minimum [10]. There are besides various other optimization methods such as evolutionary algorithm, for example, genetic algorithm, practical swarm optimization, genetic programming (GP), and so on, there are behind scope of this paper.
The principal purpose of this paper is work to experiment the achievement hybrid of SA and BP compare with BP in the FFNN structure for detection credit card fraud.
2. FEATURE SELECTION
The essential step in developing credit card fraud detection is how to extract the key features. They will influence in recognition rate and improved false alarms. By flattering feature, the data reservation will also be enhanced, so the training and time for data set will be more able for classification that runs under constant environment.
The example dataset that we are running was obtained from a data mining blog. This dataset includes the rundown of the transactions of 20,000 dynamic credit card holders recent months. The input fields incorporate credit card ID, authentication type, current balance, average bank balance, book balance, total number credit card used, and 8 distinctive cardholder classifications such as overdraft, average overdraft, number of location usage, and so on. The data set essentially gives the analysis of the cardholders’ exchanges without expressing whether the exchanges were legal or fraudulent. Concerning a given cardholder the dataset based on the following critical values, we can identify which exchange is legal or fraud:

Based on credit card usage frequency: Frequency can be found as total number card used/credit cardholder age, if the result <0.2, it implies this property is not relevant for fraud.

Based on a number of location credit card usage: Number of locations credit card used per day so far achieved from the dataset, if location is <5, it means this property is not relevant for fraud.

Based on credit card average overdraft: With respect to card used happened so far considers, the average overdraft can be found as number of overdraft/total number of card used, if overdraft with respect to card used is <0.02, it means this property is not relevant for fraud.

Based on credit card book balance: Regular book balance can be found as current book balance/average book balance, if book balance is equal or <0.25, it implies that this property is not relevant for fraud (Table I).
TABLE I Sample of Dataset
3. FFNN STRUCTURE
According to chosen features from the dataset, we created different networks. The number of hidden layers of every network is restricted to one for active and manageable calculation. The amount of neurons in the hidden layer is changed to test the results [5]. Fig. 1 illustrates the last proper structure of the network achieved it among them.
Fig. 1. Description of the FFNN developed
The log sigmoid function in equation 1 is applied as the transfer function connected by the neurons in hidden and output layers to achieve the outputs.
4. ALGORITHMS
A. Backpropagation Algorithm
BP is a common method for training ANN. The algorithm operates in two forms. First, a training input pattern is given to the input layer, which is forward to the hidden layer then into output layer to produce the network output. Mean square error (MSE) is next calculated by analyzing the estimated output and the target output for all inputs as we explained in equation 3, where “N” indicates the number of entire instances.
In the next step, besides the MSE, the network information back propagates from the output layer to the input layer, and specific connector weights are updated utilizing a “generalized ∆ rule” that is held of learning rate (η) and momentum constant (α) [11]. Equations 3 and 4 display the rule of weight updating. In particular equations, the characters “w” means the weights between the connectors “I” and “j” and “t” is the position iteration. An excellent style manual for science writers is [7].
For achieving the average η which gives the least MSE, rigorous parametric research has been conducted in this research [11]. The η at which the error is minimal is determined for the association with the SA algorithm. In this research, the momentum constant (α) is fixed equal to 0.9 for all states to speed up the learning method. The epoch size is fixed to as 1500.
B. Simulated Annulling Algorithm
The critical parameter for SA is a temperature (T) which is the similarity of the T in physical system. Beginning at a high T, the algorithm ends the minimum T with continuous decrease with attaining of a thermal equilibrium status at each T [8]. At any T, the weights are randomized. A recent set of weights is accepted as the new optimized set if the MSE with this set is under than the prior set or with a possibility that the present set of weights will reach to the global minimum. As estimated in BP, cost function and transfer functions are utilized. This research is expected if the number of adjustment in the weight set is more than 10 either the number of iterations is more than 1500 then the equilibrium state at a critical T is supposed to be done. The primary T is determined as 10°C, and ultimate T is 1°C randomly. In addition, the T is reduced with by a determinant of 0.95 random because it is challenging to obtain the accurate values of initial, ultimate, and more the threatening of T. The implementation algorithm is as follows (E(s)) is an actual function.
SA Algorithm

1) set initial solution in S

2) Set initial solution T

3) While not terminate do

4) Repeat k times

5) Chose S′ a randomly element from N(s_{i})

6) ΔE = E (S′) – E (S)

7) If (ΔE ≤ 0) then

8) S_{i+1} S′

9) else if S_{i +1} S′ with probability e^{(ΔE/T)}

10) S_{i+1} S′

11) end if

12) end repeat

13) decreased T

14) end do
C. Hybrid Algorithms
To defeat the local minimum issue of BP because of initial random weight parameters of the network, various optimization algorithms have been attempted by numerous researchers, which enhance the execution of the classification at the cost of more impalement time. In this paper I, hybridized two algorithms, joining global search SA algorithm, and local search gradient algorithm that defeats the local minimum issue with high speculation and quick union speed.
The hybrid SABP is a training algorithm joining the SA algorithm with the BP algorithm. SA is global optimization algorithm, which has a powerful capability to investigate the whole search space. This algorithm has a drawback that the search over the global optimum solution is slow. In opposite, the BP has exact and quick local searching capacity to investigate locally the optimum result, but it gets stuck to discover global optimum result in complex pursuit space.
By joining the SA and the gradientbased BP algorithm, another algorithm alluded to as hybrid SABP algorithm as shown in Fig. 2. The suggested hybrid algorithm has two stages: Initial one a global search stage, the FFNN is trained utilizing the SA algorithm for few precharacterized temperature or training error is less than some predefined value, then training mechanism changed to the second stage for searching locally utilizing a deterministic technique the BP algorithm. In this paper, it achieved SABP hybrid training algorithm as a strong option way to deal with BP algorithm.
Fig. 2. Structure of hybrid algorithm for classification
Following steps is the pseudo code for the hybrid SABP algorithm:

Randomly initialize the weights of the FFNN system appeared in Fig. 1

Evaluate weights using SA used in the neural network follow a temperature annealing schedule with the algorithm

While first temperature value is under or equal to minimum error then select the best solution for MLP then go to step 7

Select a moving method with some probability

Try a new solution

Evaluate

Select the best solution

Initialize parameters of BP learning algorithm

Initialize weights of the MLP utilizing best solution of SA

While new epoch is under or equal to maximum epoch or error converges to minimum error do

Using BP update weights to minimize error with training data

End while

Assess execution of classification with test data

End while
5. EXPERIMENTAL STUDY
The network packets that are obtained are separated into two sections. The first section about eight hundred records is used to train SA and BP neural network module. The second section is about two hundred records applied to test the credit card fraud detection. The efficiency of the neural network relies on the number, type, and amount of features and learning algorithm applied to train the neural network. Hence, as to evaluate the execution of a credit card fraud recognition strategy; we have to display a quantitative estimate. In our credit card fraud detection system, we mostly classify the network traffic into two categories, which they are normal and abnormal network traffic. Hence, we need to realize the true positive, true negative, false positive, and finally false negative to define truepositive rate (TPR) and falsenegative rate (FNR). TPR and FNR) can be calculated using the following mathematical equations [5], [6].
The TPR measures the performance of credit card fraud detection technique concerning the possibility of a suspect data reported correctly as abnormal data. Then, again the FPR measures the performance of credit card fraud detection technique as far as the possibility of a normal traffic reported as abnormal data.
As introduced, the length of parameter temperature has been taken from 10°C to 1°C. The balanced state for it includes of each 10 changes made in set of weights or 1500 iterations, the momentum is 0.9, learning rate is 0.7, maximum error to reaches is 0.01, and weight and threshold values are randomly initialized before training. Consequently, Figs. 2 and 3 show the result of training and test case for the detection rate and FPR of BP and SABP (Table II).
Fig. 3. Detection Rate of SABPFFNN and BPFFNN
TABLE II Summaries the Result During Training and Testing
Experimental result in Fig. 3. clearly show that SABPFFNN more secure in detection credit card farud in comparison to BPFFNN. Furthermore, from the Fig. 4. SABPFFNN significantly reduce the falsepositive rate compare to BPFFNN.
Fig. 4. False positive rate of SABPFFNN and BPFFNN
6. CONCLUSION
As utilization of credit cards turn out to be increasingly regular in each field of the everyday life, master card or credit card fraud has turned out to be much more rampant. To enhance security of the financial transaction frameworks in an automated and successful way, constructing an accurate and effective credit card fraud detection framework is one of the key efforts for the financial institutions. Credit card fraud detection refers to the classification and recognition issues. The paper hybrids the SA algorithm with BPFFNN for fraud detection where the simulated neural network can learn knowledge from a large number of a dataset for training and examining the result of detection. The analysis result illustrated that the using BP is a simple local minimum algorithm, and SA is a good global search algorithm or optimization algorithm based on the analysis, the experimental results indicate that the accuracy of BPFFNN is under than applied SA to BPFFNN algorithm.