You are currently viewing a new version of our website. To view the old version click .
Electronics
  • Article
  • Open Access

3 August 2018

Precision Modeling: Application of Metaheuristics on Current–Voltage Curves of Superconducting Films

,
,
,
,
,
,
and
1
Department of Electrical Engineering, COMSATS University Islamabad, Wah Campus, Wah Cantonment 47040, Pakistan
2
Department of Electrical Engineering, COMSATS University Islamabad, Abbottabad Campus, Abbottabad 22040, Pakistan
3
Department of Computer Science, COMSATS University Islamabad, Wah Campus, Wah Cantonment 47040, Pakistan
4
Department of Computer Science & IT, Ghazi University, D.G. Khan 32200, Pakistan

Abstract

Contemplating the importance of studying current–voltage curves in superconductivity, it has been recently and rightly argued that their approximation, rather than incessant measurements, seems to be a more viable option. This especially becomes bona fide when the latter needs to be recorded for a wide range of critical parameters including temperature and magnetic field, thereby becoming a tedious monotonous procedure. Artificial neural networks have been recently put forth as one methodology for approximating these so-called electrical measurements for various geometries of antidots on a superconducting thin film. In this work, we demonstrate that the prediction accuracy, in terms of mean-squared error, achieved by artificial neural networks is rather constrained, and, due to their immense credence on randomly generated networks’ coefficients, they may result in vastly varying prediction accuracies for different geometries, experimental conditions, and their own tunable parameters. This inconsistency in prediction accuracies is resolved by controlling the uncertainty in networks’ initialization and coefficients’ generation by means of a novel entropy based genetic algorithm. The proposed method helps in achieving a substantial improvement and consistency in the prediction accuracy of current–voltage curves in comparison to existing works, and is amenable to various geometries of antidots, including rectangular, square, honeycomb, and kagome, on a superconducting thin film.

1. Introduction

Precisely measuring critical current density in superconducting materials and systems requires understanding of the fundamentals of current–voltage (IV) characteristics, also called the transport measurements [1]. Recent studies have shown that these IV curves may show sudden jumps, which resemble Shapiro steps, in voltage around critical current ( I c ) and/or critical temperature ( T c ) in superconducting thin films. These steps usually appear when the vortex lattice is formed, which may lead to instability at high vortex velocities. These instabilities have been studied in different systems including nanotube Josephson Junctions [2], superconducting nanowires [3], low temperature thin films [4], and a square array of periodic antidots on an Nb film [5]. Studying such behavior in superconducting films is important since Shapiro steps are exploited to make flux Qubits that are essential for superconducting quantum computers [6]. Other mechanisms that may lead to such jumps include thermo-magnetic instabilities of vortex matter [7], thermally-assited flux flow [8], or just overheating of the superconducting film on SiO 2 substrate.
Fabrication of various geometries by electron-beam lithography, to obtain transport measurements via a physical properties measurement system (PPMS), is an expensive, tedious and cumbersome process [9]. It has been recently pointed out that these curves, especially when they are needed for a wide range of temperature and magnetic field values, may not be necessarily measured incessantly; instead, they may be obtained using some approximation technique applied on a finite amount of curves already obtained via PPMS [10,11,12]. Artificial Neural Networks (ANN) have been used as the approximation method in each of these solutions to extrapolate the IV curves for unforeseen values of critical parameters. It shall be discussed in Section 2.2 that training the ANN requires two randomly generated networks’ coefficients called weights and biases, which are updated in each iteration, until an optimal solution (with the smallest mean-squared error, MSE) is obtained. Mainly due to this randomness in the coefficient generation, ANN tends to converge to one local minima from a pool of possible solutions, which makes training of the ANN a nondeterministic process. The latter, in turn, may lead to inordinately varying values of MSE, number of iterations, and time to converge. In this way, two rounds of training on the same data may result in different prediction accuracies.
The proposed work intends to address the aforesaid problem by controlling the randomness in coefficient generation, such that the ANN always converges (close) to the global minima, by means of a novel entropy based genetic algorithm. Our case study is the prediction of IV curves for four different geometries of antidots, including rectangular, square, honeycomb, and kagome, on an Nb film. However, the proposed approach is equally applicable to other superconducting films as an alternative approximation technique for measuring other properties at the same time. The main contribution of the proposed work, therefore, is the increased accuracy in the prediction of IV curves by means of Entropy, which in a nutshell is the uncertainty measurement associated with initialization of the weights and biases. Since ANN are highly dependent on their initial conditions for their fast convergence and accurate approximation, the selected vector of weights and biases should have a minimum entropy. To the best of our knowledge, entropy, especially in conjunction with genetic algorithm (GA), has never been adopted for the prediction of IV curves for a superconducting film, which, we show, outperforms the conventionally used predictors.
The rest of the paper is organized as following: Section 2 covers our experimental setup and a brief overview of existing methodologies and relevant works. We present the problem formulation in Section 3. Section 4 presents the proposed methodology, and an overview of GA. In Section 5, we present simulation results and a comparative analysis of the used algorithms, before we conclude the paper in Section 6.

3. Problem Statement

Let Ω R ν , where ν = 2 ; ν { R × D } | R > D , be an experimental values. Let the selected feature set, ϕ = ϕ ( i ) | i Ω where ϕ 1 ( i ) , , ϕ n ( i ) ϕ are associated with the ANN training process to predict output vector ϕ ˜ p r e d . The set of features ϕ are mapped to ϕ ˜ p r e d : ϕ ϕ ˜ p r e d . The output vector ϕ ˜ p r e d is calculated as:
{ ϕ ˜ p r e d = Δ ( δ W B k , p o p t , δ R , δ s ) { 1 : 1 } , ( δ E k , δ W B , X o f s , ϕ p r e d y ) R 2 } ,
where the set of variables δ W B k , p o p t , δ R , δ s , δ E k , X o f s and ϕ p r e d y have been described in the list of symbols. The objective function in case of ANN is selected to be MSE given in the following relation:
Γ ( w ˜ ) = Ψ q = 1 Q ( Γ t a r Γ a c t ) T ( Γ t a r Γ a c t ) = Ψ q = 1 Q e q 2 .

4. Proposed Optimization of ANN’s Coefficients

4.1. Genetic Algorithms

Genetic Algorithm (GA) is an evolutionary technique belonging to a class of stochastic search algorithms, which finds an optimal solution from a pool of solutions based on the principle of survival of the fittest [18]. In GA framework, each individual is represented by a string called a chromosome, whereas a group of chromosomes generate a population. For this architecture, weights and biases vector (chromosome) is generated, which is replicated multiple times to generate a population of some pre-defined size.
Binary representation of a GA is commonly used where each chromosome is a vector c, constituting a set of m genes from the set 0 , 1 :
C i k = { G 1 k , G 2 k , , G m k } G m k { 0 , 1 } ,
where m is the length of a chromosome.
However, in practical optimization, it is more natural to represent a gene in real numbers for an optimized solution [19]. The continuous domain provides larger space and more convergence possibilities. Data range is normalized to the range of 0 1 prior to binary encoding, and, for this specific application, the chromosome is a floating point vector. In what follows, we present a few GA operators, called crossover and mutation, implemented for a thorough technical analysis.

4.1.1. Crossover Operators

Selecting a pair of chromosome C i 1 = ( G 1 1 , G 2 1 , , G m 1 ) and C i 2 = ( G 1 2 , G 2 2 , , G m 2 ) for a multiple crossover description.
  • Flat Crossover
    An offspring O i is generated where O i k = ( g 1 k , g 2 k , , g m k ) is a uniformly chosen random value from the interval [ m i n ( G i 1 , G i 2 ) , m a x ( G i 1 , G i 2 ) ] ,
    where m is the index for the number of genes and k is the index for the number of chromosomes.
  • Linear Breeder Genetic Algorithm (BGA) Crossover
    Under the same consideration as above, let O i = G 1 1 ± r a n g i . γ . Λ , where Λ = ( G i 2 G i 1 ) / G 1 G 2 and γ = l = 0 15 α l 2 l . An α { 0 , 1 } is generated randomly with the probability of p ( α i = 1 ) = 1 / 12 . Usually, r a n g i = 0.5 ( κ i τ i ) and − sign is chosen with the probability of 0.8 [20].
  • Arithmetic Crossover
    In arithmetic crossover, two Offsprings are generated, O k = ( g 1 k , g 2 k , , g m k ) , k = 1 , 2 where: g i 1 = λ G i 1 + ( 1 λ ) G i 2 and g i 2 = λ G i 2 + ( 1 λ ) G i 1 . Here, λ is a constant and user-defined value, which can vary with the number of generations.

4.1.2. Mutation Operator

Let the maximum number of generations be represented by R m a x and R t denotes the generation on which mutation operator is applied. As per mutation rule:
G i = G i + δ ( t , κ i G i ) , if γ = 0 , G i + δ ( t , G i τ i ) , if γ = 1 .

4.1.3. Selection

Selection is a process of selecting chromosomes with the smallest value of the cost function. The selection rate ς r a t e defines the survivors eligible for mating in the next generation. Generally, the ς r a t e = 50%, and the population selection is defined as:
η r e m = ς r a t e × η p o p ,
η k e e p = η p o p η r e m .
The selection probability depends on the cost weight, calculated as:
P k = C k j = 1 η k e e p C j .
The chromosome with the lowest cost in terms of mean squared error has the highest selection probability. The other selection methods include roulette wheel, rank selection and tournament [21]. Roulette wheel is a probability based method, whereas the tournament selection is a winner-takes-all based technique [22].

4.1.4. Operation of a GA

The operation of a GA is summarized as follows:
  • Create an initial population from a randomly generated weights and biases vector.
  • Repeat until the best individuals are selected:
    • Evaluate the fitness using MSE,
    • Select the parents with best fitness level,
    • Apply the selected crossover and mutation operators.
  • Terminate upon convergence.
Optimization is a process of finding the best possible solution from a given search space. In this work, an evolutionary strategy is utilized to fine-tune the parameters so as to minimize the cost function. Considering the GA’s exquisite performance on various platforms, and their ability to handle larger space problems even for stochastic objective functions, another domain is exploited, which is to optimize the solutions of artificial neural networks. The procedure has two possible directions: (1) optimize the weights and biases; and (2) optimize the architecture of ANN. In this work, the former method is used to improve the prediction accuracy.

4.2. Entropy Based GA for Optimization

For a set of discrete random variables C i k = { G 1 k , G 2 k , , G m k } , the Shannon entropy is defined as the following:
δ E = k = 0 m 1 ( G i 0 / δ G ) l o g 2 ( G i 0 / δ G ) ,
δ G = k = 0 m 1 G 0 i ,
where G i j is a vector of weights and biases called a chromosome.
The proposed cascaded design is the conjunction of three separate modules comprising an entropy calculator, GA optimizer, and ANN’s approximator. Here, the entropy calculator controls the uncertainty with the constraint of minimum randomness. The resulting vector, which is then forwarded to the GA module for optimization, comprises weights and biases with either minimum entropy or maximum repetition. Finally, the third module uses the optimized weights and biases to train the ANN for final approximation. Figure 4 summarizes the proposed methodology, and therefore the major contribution of this work, in a flow chart, and the proposed curve approximation with the optimized coefficients approach is given in Algorithm 1.
Figure 4. Flow chart depicting the proposed methodology.
Algorithm 1: Entropy based GA for weights and biases optimization
Electronics 07 00138 i001

5. Simulation Results

5.1. Design Parameters

The number of hidden layers is fixed at three, where each has 27, 17, and 8 neurons in order; this is generally represented by network’s configuration {27 17 8}. The mutation rate is defined to be 0.2 and entropy count variable κ is selected to be 200 for the optimal solution. Maximum generations of GA and maximum epochs of ANN are fixed to be 1000, while GA tolerance value ϵ is selected to be 1 × 10 8 . Learning rate δ r and initial step size δ s are selected to be 0.9 and 0.8, respectively.

5.2. Results and Discussion

Table 1 explains the numerical comparisons of five different training algorithms which include Levenberg–Marquardt (LM), three variants of Conjugate Gradient (CG), and Bayesian, utilizing three crossover techniques for GA over four different geometries. It can be clearly observed, linear-BGA crossover technique in conjunction with Bayesian regularization framework outperforms other methods in terms of minimum epochs and MSE for all geometries. For the mentioned sequence (Linear-BGA + Bayesian Regularization), epochs are in the range of (36–47), whilst minimum MSE achieved is in the range of ( 6.89 × 10 9 8.87 × 10 9 ) . The worst performance with linear-BGA is of CGP backpropagation having epochs in the range of 164–394. Similarly, with other crossover techinques, still Bayesian backpropagation acheives minimum MSE and also minimum epochs.
Table 1. Comparison of ANN’s training algorithms with multiple GA operators.
Figure 5 presents the MSE comparison of selected crossover techinques on all four geometries. It is clear from the given plots that the proposed algorithm works efficiently when the selected crossover method is linear-BGA. In order to carry out a fair comparison, the prediction for the same geometries and algorithms is also performed with conventional ANN techniques, as proposed by the benchmark solutions, as tabulated in Table 2. It may be noted that the epochs and MSE obtained by the proposed method are significantly improved, which justifies the effectiveness of our novel design in terms of prediction accuracy and consistency. In Figure 6, we have presented a comparison between the predicted and the physically measured IV curves. Note that the curve given here was not included in the training process of ANN, and was measured at 8.65 K temperature, and 41 Oe magnetic field, which are identical to the one used by the benchmark works for cross-checking the obtained result.
Figure 5. MSE comparison of various cross-over techniques on the selected geometries.
Table 2. Comparison of various ANN’s training algorithms as used in benchmark works.
Figure 6. Predicted vs. measured IV characteristics.

6. Conclusions

Being predominantly dependent upon random initial conditions, the generated network’s coefficients mostly force the artificial neural networks to converge to a different local minima on every execution. This generally leads to an inconsistent prediction accuracy, even for identical experimental setup and tunable parameters. In this work, we have proposed an entropy based genetic algorithm, and used it to control the randomness in coefficient generation. This technique forces the network to converge (close) to the global minima on every execution, which constrains the prediction accuracy, measured in mean-squared error, to a certain acceptable level. We have applied our technique to approximate the current–voltage curves for four different lattices on a superconducting film, and compared our results with three recent works, which made use of artificial neural networks to achieve the same. Our results have shown that the proposed methodology yields better consistence and greater prediction accuracy.

Author Contributions

S.R.N. conceived the entire concept, and was responsible for most of the writeup. T.A. came up with the concept of entropy controlled genetic algorithm, and was assisted by T.I. in performing MATLAB simulations. S.A.H. and H.G.U. integrated artificial neural networks with the proposed genetic algorithm, and performed MATLAB simulations. W.K. carried out an in-depth analysis of the obtained results, and made a comparison with state-of-the-art prediction techniques. A.S. proofread the manuscript, and contributed significantly in submitting the revision. M.K. performed transport measurements on Physical Properties Measurement System.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following is a list of abbreviations and symbols used throughout the paper.
R l Set of real numbers
Ω Subset of features within R l
ϕ Selected set of features
ϕ ˜ p r e d y Vector of predicted values
δ R Learning rate
δ s Step size
δ W B k Vector of weights and biases
Ψ MSE constant
δ E k Vector of entropy values
δ W B Vector of optimized weights and biases
J w e Jacobian matrix
IIdentity matrix
ϵ Error threshold
ϱ Direction set variable
r k Input vector
e q Generated error
Γ ( w ˜ ) Cost function based on mse
Γ t a r Target vector
Γ a c t Actual output vector
δ o Output layer transfer function
δ H Hidden layer transfer function
p o p t Population size
X o f s Offspring with maximum fitness
C k kth chromosome
G k kth gene
η r a t e Population selection rate
δ k e e p Selected number of individuals in population p o p t

References

  1. Acharya, S.; Bangera, K.V.; Shivakumar, G.K. Electrical characterization of vacuum-deposited p-CdTe/n-ZnSe heterojunctions. Appl. Nanosci. 2015, 5, 1003–1007. [Google Scholar] [CrossRef]
  2. Cleuziou, J.P.; Wernsdorfer, W.; Andergassen, S.; Florens, S.; Bouchiat, V.; Ondarçuhu, T.; Monthioux, M. Gate-tuned high frequency response of carbon nanotube Josephson junctions. Phys. Rev. Lett. 2007, 99, 117001. [Google Scholar] [CrossRef] [PubMed]
  3. Dinsmore, R.C., III; Bae, M.H.; Bezryadin, A. Fractional order Shapiro steps in superconducting nanowires. Appl. Phys. Lett. 2008, 93, 192505. [Google Scholar] [CrossRef]
  4. Mandal, S.; Naud, C.; Williams, O.A.; Bustarret, É.; Omnès, F.; Rodière, P.; Meunier, T.; Saminadayar, L.; Bäuerle, C. Detailed study of superconductivity in nanostructured nanocrystalline boron doped diamond thin films. Phys. Status Solidi 2010, 207, 2017–2022. [Google Scholar] [CrossRef]
  5. Kamran, M.; He, S.-K.; Zhang, W.-J.; Cao, W.-H.; Li, B.-H.; Kang, L.; Chen, J.; Wu, P.-H.; Qiu, X.-G. Matching effect in superconducting NbN thin film with a square lattice of holes. Chin. Phys. B 2009, 18, 4486. [Google Scholar] [CrossRef]
  6. Heiselberg, P. Shapiro steps in Josephson Junctions. Niels Bohr Institute, University of Copenhagen. 2013. Available online: https://cmt.nbi.ku.dk/student_projects/bsc/heiselberg.pdf (accessed on 13 July 2018).
  7. Sidorenko, A.; Zdravkov, V.; Ryazanov, V.; Horn, S.; Klimm, S.; Tidecks, R.; Wixforth, A.; Koch, T.; Schimmel, T. Thermally assisted flux flow in MgB2: Strong magnetic field dependence of the activation energy. Philos. Mag. 2005, 85, 1783–1790. [Google Scholar] [CrossRef]
  8. Fogel, N.Y.; Cherkasova, V.G.; Koretzkaya, O.A.; Sidorenko, A.S. Thermally assisted flux flow and melting transition for Mo/Si multilayers. Phys. Rev. B 1997, 55, 85. [Google Scholar] [CrossRef]
  9. Yu, C.C.; Chen, Y.T.; Wan, D.H.; Chen, H.L.; Ku, S.L.; Chou, Y.F. Using one-step, dual-side nanoimprint lithography to fabricate low-cost, highly flexible wave plates exhibiting broadband antireflection. J. Electrochem. Soc. 2011, 158, J195–J199. [Google Scholar] [CrossRef]
  10. Haider, S.A.; Naqvi, S.R.; Akram, T.; Kamran, M. Prediction of critical currents for a diluted square lattice using Artificial Neural Networks. Appl. Sci. 2017, 7, 238. [Google Scholar] [CrossRef]
  11. Haider, S.A.; Naqvi, S.R.; Akram, T.; Kamran, M.; Qadri, N.N. Modeling electrical properties for various geometries of antidots on a superconducting film. Appl. Nanosci. 2017, 7, 933–945. [Google Scholar] [CrossRef]
  12. Kamran, M.; Haider, S.A.; Akram, T.; Naqvi, S.R.; He, S.K. Prediction of IV curves for a superconducting thin film using artificial neural networks. Superlattices Microstruct. 2016, 95, 88–94. [Google Scholar] [CrossRef]
  13. Naqvi, S.R.; Akram, T.; Haider, S.A.; Kamran, M. Artificial neural networks based dynamic priority arbitration for asynchronous flow control. Neural Comput. Appl. 2016, 29, 627–637. [Google Scholar] [CrossRef]
  14. Güçlü, U.; van Gerven, M.A. Deep neural networks reveal a gradient in the complexity of neural representations across the ventral stream. J. Neurosci. 2015, 35, 10005–10014. [Google Scholar] [CrossRef] [PubMed]
  15. Haykin, S. Neural Networks and Learning Machines; Pearson: Upper Saddle River, NJ, USA, 2009; Volume 3. [Google Scholar]
  16. Zermeño, V.; Sirois, F.; Takayasu, M.; Vojenciak, M.; Kario, A.; Grilli, F. A self-consistent model for estimating the critical current of superconducting devices. Supercond. Sci. Technol. 2015, 28, 085004. [Google Scholar] [CrossRef]
  17. Bonanno, F.; Capizzi, G.; Graditi, G.; Napoli, C.; Tina, G.M. A radial basis function neural network based approach for the electrical characteristics estimation of a photovoltaic module. Appl. Energy 2012, 97, 956–961. [Google Scholar] [CrossRef]
  18. Naqvi, S.R.; Akram, T.; Iqbal, S.; Haider, S.A.; Kamran, M.; Muhammad, N. A dynamically reconfigurable logic cell: From artificial neural networks to quantum-dot cellular automata. Appl. Nanosci. 2018, 8, 89–103. [Google Scholar] [CrossRef]
  19. Haupt, R.L.; Haupt, S.E. Practical Genetic Algorithms; Wiley: New York, NY, USA, 1998; Volume 2. [Google Scholar]
  20. Herrera, F.; Lozano, M.; Sánchez, A.M. A taxonomy for the crossover operator for real-coded genetic algorithms: An experimental study. Int. J. Intell. Syst. 2003, 18, 309–338. [Google Scholar] [CrossRef]
  21. Razali, N.M.; Geraghty, J. Genetic algorithm performance with different selection strategies in solving TSP. In Proceedings of the World Congress on Engineering, London, UK, 6–8 July 2011; International Association of Engineers: Hong Kong, China, 2011; Volume 2, pp. 1134–1139. [Google Scholar]
  22. Mc Ginley, B.; Maher, J.; O’Riordan, C.; Morgan, F. Maintaining healthy population diversity using adaptive crossover, mutation, and selection. IEEE Trans. Evol. Comput. 2011, 15, 692–714. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.