Sample records for network optimization algorithms

In this paper we propose a tabu search algorithm to find the least-cost design of looped water distribution networks. The mathematical nature of this optimization problem, a nonlinear mixed integer problem, is at the origin of a multitude of contributions to the literature in the last 25 years. In fact, exact optimization methods have not been found for this type of problem, and, in the past, classical optimization methods, like linear and nonlinear programming, were tried at the cost of dras...

This paper examines the forecasting performance of multi-layer feed forward neural networks in modeling a particular foreign exchange rates, i.e. Japanese Yen/US Dollar. The effects of two learning methods, Back Propagation and Genetic Algorithm, in which the neural network topology and other parameters fixed, were investigated. The early results indicate that the application of this hybrid system seems to be well suited for the forecasting of foreign exchange rates. The Neural Networks and Genetic Algorithm were programmed using MATLAB«.

Full Text Available Various approaches and methods are used for designing of optimum deployment of Passive Optical Networks (PON according to selected optimization criteria, such as optimal trenching distance, endpoint attenuation and overall installed fibre length. This article describes the ideas and possibilities for an algorithm with the application of graph algorithms for finding the shortest path from Optical Line Termination to Optical Network Terminal unit. This algorithm uses a combination of different methods for generating of an optimal metric, thus creating the optimized tree topology mainly focused on summary trenching distance. Furthermore, it deals with algorithms for finding an optimal placement of optical splitter with the help of K-Means clustering method and hierarchical clustering technique. The results of the proposed algorithm are compared with existing methods.

Full Text Available When confronting the complex problems, radial basis function (RBF neural network has the advantages of adaptive and self-learning ability, but it is difficult to determine the number of hidden layer neurons, and the weights learning ability from hidden layer to the output layer is low; these deficiencies easily lead to decreasing learning ability and recognition precision. Aiming at this problem, we propose a new optimized RBF neural networkalgorithm based on genetic algorithm (GA-RBF algorithm, which uses genetic algorithm to optimize the weights and structure of RBF neural network; it chooses new ways of hybrid encoding and optimizing simultaneously. Using the binary encoding encodes the number of the hidden layer’s neurons and using real encoding encodes the connection weights. Hidden layer neurons number and connection weights are optimized simultaneously in the new algorithm. However, the connection weights optimization is not complete; we need to use least mean square (LMS algorithm for further leaning, and finally get a new algorithm model. Using two UCI standard data sets to test the new algorithm, the results show that the new algorithm improves the operating efficiency in dealing with complex problems and also improves the recognition precision, which proves that the new algorithm is valid.

Full Text Available Delay Tolerant Networks (DTNs are an emerging type of networks which do not need a predefined infrastructure. In fact, data forwarding in DTNs relies on the contacts among nodes which may possess different features, radio range, battery consumption and radio interfaces. On the other hand, efficient message delivery under limited resources, e.g., battery or storage, requires to optimize forwarding policies. We tackle optimal forwarding control for a DTN composed of nodes of different types, forming a so-called heterogeneous network. Using our model, we characterize the optimal policies and provide a suitable framework to design a new class of multi-dimensional stochastic approximation algorithms working for heterogeneous DTNs. Crucially, our proposed algorithms drive online the source node to the optimal operating point without requiring explicit estimation of network parameters. A thorough analysis of the convergence properties and stability of our algorithms is presented.

and the pipe friction and heat loss formulations are non-linear. In order to find the optimal district heating network configuration, genetic algorithm which handles the mixed integer nonlinear programming problem is chosen. The network configuration is represented with binary and integer encoding...

Convolutional neural network (CNN) is one of the most prominent architectures and algorithm in Deep Learning. It shows a remarkable improvement in the recognition and classification of objects. This method has also been proven to be very effective in a variety of computer vision and machine learning problems. As in other deep learning, however, training the CNN is interesting yet challenging. Recently, some metaheuristic algorithms have been used to optimize CNN using Genetic Algorithm, Parti...

In this letter, aiming to obtain the best multicast performance of optical network in which the video conference information is carried by specified wavelength, we extend the solutions of matrix games with the network coding theory and devise a new method to solve the complex problems of multicast network switching. In addition, an experimental optical network has been testified with best switching strategies by employing the novel numerical solution designed with an effective way of genetic algorithm. The result shows that optimal solutions with genetic algorithm are accordance with the ones with the traditional fictitious play method.

Full Text Available Multiconstraints optimalnetwork load balancing is an NP-hard problem and it is an important part of traffic engineering. In this research we balance the network load using classical method (brute force approach and dynamic programming is used but result shows the limitation of this method but at a certain level we recognized that the optimization of balanced network load with increased number of nodes and demands is intractable using the classical method because the solution set increases exponentially. In such case the optimization techniques like evolutionary techniques can employ for optimizingnetwork load balance. In this paper we analyzed proposed classical algorithm and evolutionary based genetic approach is devise as well as proposed in this paper for optimizing the balance network load.

A geodetic network is a network which is measured exactly by techniques of terrestrial surveying based on measurement of angles and distances and can control stability of dams, towers and their around lands and can monitor deformation of surfaces. The main goals of an optimal geodetic network design process include finding proper location of control station (First order Design) as well as proper weight of observations (second order observation) in a way that satisfy all the criteria considered for quality of the network with itself is evaluated by the network's accuracy, reliability (internal and external), sensitivity and cost. The first-order design problem, can be dealt with as a numeric optimization problem. In this designing finding unknown coordinates of network stations is an important issue. For finding these unknown values, network geodetic observations that are angle and distance measurements must be entered in an adjustment method. In this regard, using inverse problem algorithms is needed. Inverse problem algorithms are methods to find optimal solutions for given problems and include classical and evolutionary computations. The classical approaches are analytical methods and are useful in finding the optimum solution of a continuous and differentiable function. Least squares (LS) method is one of the classical techniques that derive estimates for stochastic variables and their distribution parameters from observed samples. The evolutionary algorithms are adaptive procedures of optimization and search that find solutions to problems inspired by the mechanisms of natural evolution. These methods generate new points in the search space by applying operators to current points and statistically moving toward more optimal places in the search space. Genetic algorithm (GA) is an evolutionary algorithm considered in this paper. This algorithm starts with definition of initial population, and then the operators of selection, replication and variation are applied

Marquardt algorithm by varying conditions such as inputs, hidden neurons, initialization, training sets and random Gaussian noise injection to ... Several such ensembles formed the population which was evolved to generate the fittest ensemble.

Full Text Available In order to serve for the ecological monitoring efficiency of Poyang Lake, an improved hybrid algorithm, mixed with differential evolution and particle swarm optimization, is proposed and applied to optimize the coverage problem of wireless sensor network. And then, the affect of the population size and the number of iterations on the coverage performance are both discussed and analyzed. The four kinds of statistical results about the coverage rate are obtained through lots of simulation experiments.

the heating plant location is allowed to vary. The connection between the heat generation plant and the end users can be represented with mixed integer and the pipe friction and heat loss formulations are non-linear. In order to find the optimal DH distribution pipeline configuration, the genetic algorithm...... by multi factors as the consumer heating load, the distance between the heating plant to the consumer, the design criteria regarding pressure and temperature limitation, as well as the corresponding network heat loss....

Full Text Available The advantages of neural network technology is shown in comparison of traditional descriptions of dynamically changing systems, which include a modern land market. The basic difficulty arising in the practical implementation of neural network models of the land market and construction products is revealed It is the formation of a representative set of training and test examples. The requirements which are necessary for the correct description of the current economic situation has been determined, it consists in the fact that Train-paid-set in the feature space should not has the ranges with a low density of observations. The methods of optimization of empirical array, which allow to avoid the long-range extrapolation of data from range of concentration of the set of examples are formulated. It is shown that a radical method of optimization a set of training and test examples enclosing to collect supplemantary information, is associated with significant costs time and resources for the economic problems and the ratio of cost / efficiency is less efficient than an algorithmoptimization neural network models the earth market fixed set of empirical data. Algorithm of optimization based on the transformation of arrays of information which represents the expansion of the ranges of concentration of the set of examples and compression the ranges of low density of observations is analyzed in details. The significant reduction in the relative error of land price description is demonstrated on the specific example of Voronezh region market of lands which intend for road construction, it makes the using of radical method of empirical optimization of the array costeffective with accounting the significant absolute value of the land. The high economic efficiency of the proposed algorithms is demonstrated.

In order to reduce the energy consumption of wireless sensor network and improve the survival time of network, the clustering routing protocol of wireless sensor networks based on chicken swarm optimizationalgorithm was proposed. On the basis of LEACH agreement, it was improved and perfected that the points on the cluster and the selection of cluster head using the chicken group optimizationalgorithm, and update the location of chicken which fall into the local optimum by Levy flight, enhance population diversity, ensure the global search capability of the algorithm. The new protocol avoided the die of partial node of intensive using by making balanced use of the network nodes, improved the survival time of wireless sensor network. The simulation experiments proved that the protocol is better than LEACH protocol on energy consumption, also is better than that of clustering routing protocol based on particle swarm optimizationalgorithm.

Full Text Available Artificial Neural Network (ANN design is a complex task because its performance depends on the architecture, the selected transfer function, and the learning algorithm used to train the set of synaptic weights. In this paper we present a methodology that automatically designs an ANN using particle swarm optimizationalgorithms such as Basic Particle Swarm Optimization (PSO, Second Generation of Particle Swarm Optimization (SGPSO, and a New Model of PSO called NMPSO. The aim of these algorithms is to evolve, at the same time, the three principal components of an ANN: the set of synaptic weights, the connections or architecture, and the transfer functions for each neuron. Eight different fitness functions were proposed to evaluate the fitness of each solution and find the best design. These functions are based on the mean square error (MSE and the classification error (CER and implement a strategy to avoid overtraining and to reduce the number of connections in the ANN. In addition, the ANN designed with the proposed methodology is compared with those designed manually using the well-known Back-Propagation and Levenberg-Marquardt Learning Algorithms. Finally, the accuracy of the method is tested with different nonlinear pattern classification problems.

This paper presents a novel optimizationalgorithm, namely, hierarchical artificial bee colony optimization, called HABC, to tackle the radio frequency identification network planning (RNP) problem. In the proposed multilevel model, the higher-level species can be aggregated by the subpopulations from lower level. In the bottom level, each subpopulation employing the canonical ABC method searches the part-dimensional optimum in parallel, which can be constructed into a complete solution for the upper level. At the same time, the comprehensive learning method with crossover and mutation operators is applied to enhance the global search ability between species. Experiments are conducted on a set of 10 benchmark optimization problems. The results demonstrate that the proposed HABC obtains remarkable performance on most chosen benchmark functions when compared to several successful swarm intelligence and evolutionary algorithms. Then HABC is used for solving the real-world RNP problem on two instances with different scales. Simulation results show that the proposed algorithm is superior for solving RNP, in terms of optimization accuracy and computation robustness.

Full Text Available This paper presents a novel optimizationalgorithm, namely, hierarchical artificial bee colony optimization, called HABC, to tackle the radio frequency identification network planning (RNP problem. In the proposed multilevel model, the higher-level species can be aggregated by the subpopulations from lower level. In the bottom level, each subpopulation employing the canonical ABC method searches the part-dimensional optimum in parallel, which can be constructed into a complete solution for the upper level. At the same time, the comprehensive learning method with crossover and mutation operators is applied to enhance the global search ability between species. Experiments are conducted on a set of 10 benchmark optimization problems. The results demonstrate that the proposed HABC obtains remarkable performance on most chosen benchmark functions when compared to several successful swarm intelligence and evolutionary algorithms. Then HABC is used for solving the real-world RNP problem on two instances with different scales. Simulation results show that the proposed algorithm is superior for solving RNP, in terms of optimization accuracy and computation robustness.

This paper will consider the case for genetic algorithmoptimization in the development of an artificial neural network model. It will provide a methodological evaluation of reported investigations with respect to hydrological forecasting and prediction. The intention in such operations is to develop a superior modelling solution that will be: \\begin{itemize} more accurate in terms of output precision and model estimation skill; more tractable in terms of personal requirements and end-user control; and/or more robust in terms of conceptual and mechanical power with respect to adverse conditions. The genetic algorithmoptimization toolbox could be used to perform a number of specific roles or purposes and it is the harmonious and supportive relationship between neural networks and genetic algorithms that will be highlighted and assessed. There are several neural network mechanisms and procedures that could be enhanced and potential benefits are possible at different stages in the design and construction of an operational hydrological model e.g. division of inputs; identification of structure; initialization of connection weights; calibration of connection weights; breeding operations between successful models; and output fusion associated with the development of ensemble solutions. Each set of opportunities will be discussed and evaluated. Two strategic questions will also be considered: [i] should optimization be conducted as a set of small individual procedures or as one large holistic operation; [ii] what specific function or set of weighted vectors should be optimized in a complex software product e.g. timings, volumes, or quintessential hydrological attributes related to the 'problem situation' - that might require the development flood forecasting, drought estimation, or record infilling applications. The paper will conclude with a consideration of hydrological forecasting solutions developed on the combined methodologies of co-operative co-evolution and

An industrial case study in the form of a large-scale hydraulic network underlying a district heating system is considered. A distributed control is developed that minimizes the aggregated electrical energy consumption of the pumps in the network without violating the control demands. The algorithm...... a Plug & Play control system as most commissioning can be done during the manufacture of the pumps. Only information on the graph-structure of the hydraulic network is needed during installation....

This paper researches the drainage routing problem in drainage pipe network, and propose an intelligent scheduling method. The method relates to the design of improved particle swarm optimizationalgorithm, the establishment of the corresponding model from the pipe network, and the process by using the algorithm based on improved particle swarm optimization to find the optimum drainage route in the current environment.

Most thermal processes encountered in the power industry (chemical, metallurgical, nuclear and thermal power stations) necessitate the transfer of large amounts of heat between fluids having different thermal potentials. A common practice applied to achieve such a requirement consists of using heat exchangers. In general, each current of fluid is conveniently cooled or heated independently from each other in the power plant. When the number of heat exchangers is large enough, however, a convenient arrangement of different flow currents may allow a considerable reduction in energy consumption to be obtained (Linnhoff and Hidmarsh, 1983). In such a case the heat exchangers form a 'Heat Exchanger Network' (HEN) that can be optimized to reduce the overall energy consumption. This type of optimization problem, involves two separates calculation procedures. First, it is necessary to optimize the topology of the HEN that will permit a reduction in energy consumption to be obtained. In a second step the power distribution across the HEN should be optimized without violating the second law of thermodynamics. The numerical treatment of this kind of problem requires the use of both discrete variables (for taking into account each heat exchanger unit) and continuous variables for handling the thermal load of each unit. It is obvious that for a large number of heat exchangers, the use of conventional calculation methods, i.e., Simplexe, becomes almost impossible. Therefore, in this paper we present a 'Genetic Algorithm' (GA), that has been implemented and successfully used to treat complex HENs, containing a large number of heat exchangers. As opposed to conventional optimization techniques that require the knowledge of the derivatives of a function, GAs start the calculation process from a large population of possible solutions of a given problem (Goldberg, 1999). Each possible solution is in turns evaluated according to a 'fitness' criterion obtained from an objective

Full Text Available In present days, continuous effort is being made in bringing down the line losses of the electrical distribution networks. Therefore proper allocation of capacitors is of utmost importance because, it will help in reducing the line losses and maintaining the bus voltage. This in turn results in improving the stability and reliability of the system. In this paper Whale OptimizationAlgorithm (WOA is used to find optimal sizing and placement of capacitors for a typical radial distribution system. Multi objectives such as operating cost reduction and power loss minimization with inequality constraints on voltage limits are considered and the proposed algorithm is validated by applying it on standard radial systems: IEEE-34 bus and IEEE-85 bus radial distribution test systems. The results obtained are compared with those of existing algorithms. The results show that the proposed algorithm is more effective in bringing down the operating costs and in maintaining better voltage profile. Keywords: Whale OptimizationAlgorithm (WOA, Optimal allocation and sizing of capacitors, Power loss reduction and voltage stability improvement, Radial distribution system, Operating cost minimization

Full Text Available This study proposes a traffic congestion minimization model in which the traffic signal setting optimization is performed through a combined simulation-optimization model. In this model, the TRANSYT traffic simulation software is combined with Differential Evolution (DE optimizationalgorithm, which is based on the natural selection paradigm. In this context, the EQuilibrium Network Design (EQND problem is formulated as a bilevel programming problem in which the upper level is the minimization of the total network performance index. In the lower level, the traffic assignment problem, which represents the route choice behavior of the road users, is solved using the Path Flow Estimator (PFE as a stochastic user equilibrium assessment. The solution of the bilevel EQND problem is carried out by the proposed Differential Evolution and TRANSYT with PFE, the so-called DETRANSPFE model, on a well-known signal controlled test network. Performance of the proposed model is compared to that of two previous works where the EQND problem has been solved by Genetic-Algorithms- (GAs- and Harmony-Search- (HS- based models. Results show that the DETRANSPFE model outperforms the GA- and HS-based models in terms of the network performance index and the computational time required.

Full Text Available In this paper, constrained and unconstrained versions of a new formulation of Ant Colony OptimizationAlgorithm (ACOA named Arc Based Ant Colony OptimizationAlgorithm (ABACOA are augmented with the Tree Growing Algorithm (TGA and used for the optimal layout and pipe size design of gravitational sewer networks. The main advantages offered by the proposed ABACOA formulation are proper definition of heuristic information, a useful component of the ant-based algorithms, and proper trade-off between the two conflicting search attributes of exploration and exploitation. In both the formulations, the TGA is used to incrementally construct feasible tree-like layouts out of the base layout. In the first formulation, unconstrained version of ABACOA is used to determine the nodal cover depths of sewer pipes while in the second formulation, a constrained version of ABACOA is used to determine the nodal cover depths of sewer pipes which satisfy the pipe slopes constraint. Three different methods of cut determination are also proposed to complete the construction of a tree-like network containing all base layout pipes, here. The proposed formulations are used to solve three test examples of different scales and the results are presented and compared with other available results in the literature. Comparison of the results shows that best results are obtained using the third cutting method in both the formulations. In addition, the results indicate the ability of the proposed methods and in particular the constrained version of ABACOA equipped with TGA to solve sewer networks design optimization problem. To be specific, the constrained version of ABACOA has been able to produce results 0.1%, 1% and 2.1% cheaper than those obtained by the unconstrained version of ABACOA for the first, second and the third test examples, respectively.

The article discusses the utilization of a modified genetic algorithm (GA) for the optimization of the shunt compensation in meshed and radial MV distribution networks. The algorithm looks for minimum costs of the network power losses and minimum capital and operating costs of applied capacitors...

The infrastructure networks, including the Internet, telecommunication networks, electrical power grids, transportation networks (road, railway, waterway, and airway networks), gas networks and water networks, are becoming more and more complex. The complex infrastructure networks are crucial to our

Full Text Available This paper presents an application of Chaotic differential evolution optimization approach meta-heuristics in solving transmission network expansion planning TNEP using an AC model associated with reactive power planning RPP. The reliabilityredundancy of network analysis optimization problems implicate selection of components with multiple choices and redundancy levels that produce maximum benefits can be subject to the cost weight and volume constraints is presented in this paper. Classical mathematical methods have failed in handling non-convexities and non-smoothness in optimization problems. As an alternative to the classical optimization approaches the meta-heuristics have attracted lot of attention due to their ability to find an almost global optimal solution in reliabilityredundancy optimization problems. Evolutionary algorithms EAs paradigms of evolutionary computation field are stochastic and robust meta-heuristics useful to solve reliabilityredundancy optimization problems. EAs such as genetic algorithm evolutionary programming evolution strategies and differential evolution are being used to find global or near global optimal solution. The Differential Evolution Algorithm DEA population-based algorithm is an optimalalgorithm with powerful global searching capability but it is usually in low convergence speed and presents bad searching capability in the later evolution stage. A new Chaotic Differential Evolution algorithm CDE based on the cat map is recommended which combines DE and chaotic searching algorithm. Simulation results and comparisons show that the chaotic differential evolution algorithm using Cat map is competitive and stable in performance with other optimization approaches and other maps.

The problem of environmental monitoring networksoptimization (MNO) belongs to one of the basic and fundamental tasks in spatio-temporal data collection, analysis, and modeling. There are several approaches to this problem, which can be considered as a design or redesign of monitoring network by applying some optimization criteria. The most developed and widespread methods are based on geostatistics (family of kriging models, conditional stochastic simulations). In geostatistics the variance is mainly used as an optimization criterion which has some advantages and drawbacks. In the present research we study an application of advanced techniques following from the statistical learning theory (SLT) - support vector machines (SVM) and the optimization of monitoring networks when dealing with a classification problem (data are discrete values/classes: hydrogeological units, soil types, pollution decision levels, etc.) is considered. SVM is a universal nonlinear modeling tool for classification problems in high dimensional spaces. The SVM solution is maximizing the decision boundary between classes and has a good generalization property for noisy data. The sparse solution of SVM is based on support vectors - data which contribute to the solution with nonzero weights. Fundamentally the MNO for classification problems can be considered as a task of selecting new measurement points which increase the quality of spatial classification and reduce the testing error (error on new independent measurements). In SLT this is a typical problem of active learning - a selection of the new unlabelled points which efficiently reduce the testing error. A classical approach (margin sampling) to active learning is to sample the points closest to the classification boundary. This solution is suboptimal when points (or generally the dataset) are redundant for the same class. In the present research we propose and study two new advanced methods of active learning adapted to the solution of

a hybrid optimization method that aims of maximizing the Net Present Value related to the Investment made by Wind Turbines developers in an active distribution network. The proposed network combines a Genetic Algorithm with a multi-period optimal power flow. The method, integrating active management...

Full Text Available A new energy efficient optimal Connected Dominating Set (CDS algorithm with activity scheduling for mobile ad hoc networks (MANETs is proposed. This algorithm achieves energy efficiency by minimizing the Broadcast Storm Problem [BSP] and at the same time considering the node’s remaining energy. The Connected Dominating Set is widely used as a virtual backbone or spine in mobile ad hoc networks [MANETs] or Wireless Sensor Networks [WSN]. The CDS of a graph representing a network has a significant impact on an efficient design of routing protocol in wireless networks. Here the CDS is a distributed algorithm with activity scheduling based on unit disk graph [UDG]. The node’s mobility and residual energy (RE are considered as parameters in the construction of stable optimal energy efficient CDS. The performance is evaluated at various node densities, various transmission ranges, and mobility rates. The theoretical analysis and simulation results of this algorithm are also presented which yield better results.

A new energy efficient optimal Connected Dominating Set (CDS) algorithm with activity scheduling for mobile ad hoc networks (MANETs) is proposed. This algorithm achieves energy efficiency by minimizing the Broadcast Storm Problem [BSP] and at the same time considering the node's remaining energy. The Connected Dominating Set is widely used as a virtual backbone or spine in mobile ad hoc networks [MANETs] or Wireless Sensor Networks [WSN]. The CDS of a graph representing a network has a significant impact on an efficient design of routing protocol in wireless networks. Here the CDS is a distributed algorithm with activity scheduling based on unit disk graph [UDG]. The node's mobility and residual energy (RE) are considered as parameters in the construction of stable optimal energy efficient CDS. The performance is evaluated at various node densities, various transmission ranges, and mobility rates. The theoretical analysis and simulation results of this algorithm are also presented which yield better results.

This paper addresses weight optimization problem in distributed consensus averaging algorithm over networks with symmetric star topology. We have determined optimal weights and convergence rate of the network in terms of its topological parameters. In addition, two alternative topologies with more rapid convergence rates have been introduced. The new topologies are Complete-Cored Symmetric (CCS) star and K-Cored Symmetric (KCS) star topologies. It has been shown that the optimal weights for t...

Full Text Available Optimal allocation of generalized power sources in distribution network is researched. A simple index of voltage stability is put forward. Considering the investment and operation benefit, the stability of voltage and the pollution emissions of generalized power sources in distribution network, a multi-objective optimization planning model is established. A multi-objective particle swarm optimizationalgorithm is proposed to solve the optimal model. In order to improve the global search ability, the strategies of fast non-dominated sorting, elitism and crowding distance are adopted in this algorithm. Finally, tested the model and algorithm by IEEE-33 node system to find the best configuration of GP, the computed result shows that with the generalized power reasonable access to the active distribution network, the investment benefit and the voltage stability of the system is improved, and the proposed algorithm has better global search capability.

Full Text Available Integration of Renewable Distributed Generations (RDGs such as photovoltaic (PV systems and wind turbines (WTs in distribution networks can be considered a brilliant and efficient solution to the growing demand for energy. This article introduces new robust and effective techniques like hybrid Particle Swarm Optimization in addition to a Gravitational Search Algorithm (PSOGSA and Moth-Flame Optimization (MFO that are proposed to deduce the optimum location with convenient capacity of RDGs units for minimizing system power losses and operating cost while improving voltage profile and voltage stability. This paper describes two stages. First, the Loss Sensitivity Factors (LSFs are employed to select the most candidate buses for RDGs location. In the second stage, the PSOGSA and MFO are implemented to deduce the optimal location and capacity of RDGs from the elected buses. The proposed schemes have been applied on 33-bus and 69-bus IEEE standard radial distribution systems. To insure the suggested approaches validity, the numerical results have been compared with other techniques like Backtracking Search OptimizationAlgorithm (BSOA, Genetic Algorithm (GA, Particle Swarm Algorithm (PSO, Novel combined Genetic Algorithm and Particle Swarm Optimization (GA/PSO, Simulation Annealing Algorithm (SA, and Bacterial Foraging OptimizationAlgorithm (BFOA. The evaluated results have been confirmed the superiority with high performance of the proposed MFO technique to find the optimal solutions of RDGs units’ allocation. In this regard, the MFO is chosen to solve the problems of Egyptian Middle East distribution network as a practical case study with the optimal integration of RDGs.

Full Text Available Heterogeneous wireless networks are an integration of two different networks. For better performance, connections are to be exchanged among the different networks using seamless Vertical Handoff. The evolutionary algorithm of invasive weed optimizationalgorithm popularly known as the IWO has been used in this paper, to solve the Vertical Handoff (VHO and Horizontal Handoff (HHO problems. This integer coded algorithm is based on the colonizing behavior of weed plants and has been developed to optimize the system load and reduce the battery power consumption of the Mobile Node (MN. Constraints such as Receiver Signal Strength (RSS, battery lifetime, mobility, load and so on are taken into account. Individual as well as a combination of a number of factors are considered during decision process to make it more effective. This paper brings out the novel method of IWO algorithm for decision making during Vertical Handoff. Therefore the proposed VHO decision making algorithm is compared with the existing SSF and OPTG methods.

This work presents a prototype of a system for nuclear reactor core design optimization based on genetic algorithms and artificial neural networks. A neural network is modeled and trained in order to predict the flux and the neutron multiplication factor values based in the enrichment, network pitch and cladding thickness, with average error less than 2%. The values predicted by the neural network are used by a genetic algorithm in this heuristic search, guided by an objective function that rewards the high flux values and penalizes multiplication factors far from the required value. Associating the quick prediction - that may substitute the reactor physics calculation code - with the global optimization capacity of the genetic algorithm, it was obtained a quick and effective system for nuclear reactor core design optimization. (author). 11 refs., 8 figs., 3 tabs

Full Text Available In this paper, we proposed and developed Function-Oriented Networking (FON, a platform for network users. It has a different philosophy as opposed to technologies for network managers of Software-Defined Networking technology, OpenFlow. It is a technology that can immediately reflect the demands of the network users in the network, unlike the existing OpenFlow and Network Functions Virtualization (NFV, which do not reflect directly the needs of the network users. It allows the network user to determine the policy of the direct network, so it can be applied more precisely than the policy applied by the network manager. This is expected to increase the satisfaction of the service users when the network users try to provide new services. We developed FON function that performs on-demand routing for Low-Delay Required service. We analyzed the characteristics of the Ant Colony Optimization (ACO algorithm and found that the algorithm is suitable for low-delay required services. It was also the first in the world to implement the routing software using ACO Algorithm in the real Ethernet network. In order to improve the routing performance, several algorithms of the ACO Algorithm have been developed to enable faster path search-routing and path recovery. The relationship between the network performance index and the ACO routing parameters is derived, and the results are compared and analyzed. Through this, it was possible to develop the ACO algorithm.

The article discusses the utilization of a modified genetic algorithm (GA) for the optimization of the shunt compensation in meshed and radial MV distribution networks. The algorithm looks for minimum costs of the network power losses and minimum capital and operating costs of applied capacitors...... on the steady-state analysis using a linear one-line diagram model of a power network. The results of steady-state solutions are compared with the results from the DIgSILENT PowerFactory program. Its practical applicability is demonstrated on examples of 22 kV and meshed overhead distribution networks....

Full Text Available Wireless Sensor Network (WSN is a network which formed with a maximum number of sensor nodes which are positioned in an application environment to monitor the physical entities in a target area, for example, temperature monitoring environment, water level, monitoring pressure, and health care, and various military applications. Mostly sensor nodes are equipped with self-supported battery power through which they can perform adequate operations and communication among neighboring nodes. Maximizing the lifetime of the Wireless Sensor networks, energy conservation measures are essential for improving the performance of WSNs. This paper proposes an Enhanced PSO-Based Clustering Energy Optimization (EPSO-CEO algorithm for Wireless Sensor Network in which clustering and clustering head selection are done by using Particle Swarm Optimization (PSO algorithm with respect to minimizing the power consumption in WSN. The performance metrics are evaluated and results are compared with competitive clustering algorithm to validate the reduction in energy consumption.

Wireless Sensor Network (WSN) is a network which formed with a maximum number of sensor nodes which are positioned in an application environment to monitor the physical entities in a target area, for example, temperature monitoring environment, water level, monitoring pressure, and health care, and various military applications. Mostly sensor nodes are equipped with self-supported battery power through which they can perform adequate operations and communication among neighboring nodes. Maximizing the lifetime of the Wireless Sensor networks, energy conservation measures are essential for improving the performance of WSNs. This paper proposes an Enhanced PSO-Based Clustering Energy Optimization (EPSO-CEO) algorithm for Wireless Sensor Network in which clustering and clustering head selection are done by using Particle Swarm Optimization (PSO) algorithm with respect to minimizing the power consumption in WSN. The performance metrics are evaluated and results are compared with competitive clustering algorithm to validate the reduction in energy consumption.

Full Text Available The field of processing information provided by measurement results is one of the most important components of geodetic technologies. The dynamic development of this field improves classic algorithms for numerical calculations in the aspect of analytical solutions that are difficult to achieve. Algorithms based on artificial intelligence in the form of artificial neural networks, including the topology of connections between neurons have become an important instrument connected to the problem of processing and modelling processes. This concept results from the integration of neural networks and parameter optimization methods and makes it possible to avoid the necessity to arbitrarily define the structure of a network. This kind of extension of the training process is exemplified by the algorithm called the Group Method of Data Handling (GMDH, which belongs to the class of evolutionary algorithms. The article presents a GMDH type network, used for modelling deformations of the geometrical axis of a steel chimney during its operation.

Full Text Available The increasing penetration of distributed energy resources in distribution systems has brought a number of network management and operational challenges; reactive power variation has been identified as one of the dominant effects. Enormous growth in a variety of controllable devices that have complex control requirements are integrated in distribution networks. The operation modes of traditional centralized control are difficult to tackle these problems with central controller. When considering the non-linear multi-objective functions with discrete and continuous optimization variables, the proposed random gradient-free algorithm is employed to the optimal operation of controllable devices for reactive power optimization. This paper presents a distributed reactive power optimizationalgorithm that can obtain the global optimum solution based on random gradient-free algorithm for distribution network without requiring a central coordinator. By utilizing local measurements and local communications among capacitor banks and distributed generators (DGs, the proposed reactive power control strategy can realize the overall network voltage optimization and power loss minimization simultaneously. Simulation studies on the modified IEEE-69 bus distribution systems demonstrate the effectiveness and superiority of the proposed reactive power optimization strategy.

Air quality monitoring networks play a significant role in identifying the spatiotemporal patterns of air pollution, and they need to be deployed efficiently, with a minimum number of sites. The revision and optimal adjustment of existing monitoring networks is crucial for cities that have undergone rapid urban expansion and experience temporal variations in pollution patterns. The approach based on the Weather Research and Forecasting-California PUFF (WRF-CALPUFF) model and genetic algorithm (GA) was developed to design an optimal monitoring network. The maximization of coverage with minimum overlap and the ability to detect violations of standards were developed as the design objectives for redistributed networks. The non-dominated sorting genetic algorithm was applied to optimize the network size and site locations simultaneously for Shijiazhuang city, one of the most polluted cities in China. The assessment on the current network identified the insufficient spatial coverage of SO2 and NO2 monitoring for the expanding city. The optimization results showed that significant improvements were achieved in multiple objectives by redistributing the original network. Efficient coverage of the resulting designs improved to 60.99% and 76.06% of the urban area for SO2 and NO2, respectively. The redistributing design for multi-pollutant including 8 sites was also proposed, with the spatial representation covered 52.30% of the urban area and the overlapped areas decreased by 85.87% compared with the original network. The abilities to detect violations of standards were not improved as much as the other two objectives due to the conflicting nature between the multiple objectives. Additionally, the results demonstrated that the algorithm was slightly sensitive to the parameter settings, with the number of generations presented the most significant effect. Overall, our study presents an effective and feasible procedure for air quality networkoptimization at a city scale.

The deployment of passive optical networks (PONs) is slow today, especially in Europe, because completely new optical infrastructures are necessary to be installed in the last-mile segments of access networks, which is always very expensive process. One of the possibilities is to design economically effective topologies and to optimize the deployment cost. This article describes the method leading to evaluate an algorithm for designing suboptimal economic solutions and topologies for PONs by ...

Hybrid wireless-optical broadband access network (WOBAN) or Fiber-Wireless (FiWi) is the integration of wireless access network and optical network. This hybrid multi-domain network adopts the advantages of wireless and optical domains and serves the demand of technology savvy users. FiWi exhibits the properties of cost effectiveness, robustness, flexibility, high capacity, reliability and is self organized. Optical Network Unit (ONU) placement problem in FiWi contributes in simplifying the network design and enhances the performance in terms of cost efficiency and increased throughput. Several individual-based algorithms, such as Simulated Annealing (SA), Tabu Search, etc. have been suggested for ONU placement, but these algorithms suffer from premature convergence (trapping in a local optima). The present research work undertakes the deployment of FiWi and proposes a novel nature-inspired heuristic paradigm called Moth-Flame optimization (MFO) algorithm for multiple optical network units' placement. MFO is a population based algorithm. Population-based algorithms are better in handling local optima avoidance. The simulation results are compared with the existing Greedy and Simulated Annealing algorithms to optimize the position of ONUs. To the best of our knowledge, MFO algorithm has been used for the first time in this domain, moreover it has been able to provide very promising and competitive results. The performance of MFO algorithm has been analyzed by varying the 'b' parameter. MFO algorithm results in faster convergence than the existing strategies of Greedy and SA and returns a lower value of overall cost function. The results exhibit the dependence of the objective function on the distribution of wireless users also.

Full Text Available Taking into consideration both shipping and pipeline transport, this paper first analysed the risk factors for different modes of crude oil import transportation. Then, based on the minimum of both transportation cost and overall risk, a multi-objective programming model was established to optimize the transportation network of crude oil import, and the genetic algorithm and ant colony algorithm were employed to solve the problem. The optimized result shows that VLCC (Very Large Crude Carrier is superior in long distance sea transportation, whereas pipeline transport is more secure than sea transport. Finally, this paper provides related safeguard suggestions on crude oil import transportation.

Full Text Available Capacitated single allocation hub-and-spoke networks can be abstracted as a mixed integer linear programming model equation with three variables. Introducing an improved ant colony algorithm, which has six local search operators. Meanwhile, introducing the "Solution Pair" concept to decompose and optimize the composition of the problem, the problem can become more specific and effectively meet the premise and advantages of using ant colony algorithm. Finally, location simulation experiment is made according to Australia Post data to demonstrate this algorithm has good efficiency and stability for solving this problem.

Full Text Available Recent advances in modern manufacturing industries have created a great need to track and identify objects and parts by obtaining real-time information. One of the main technologies which has been utilized for this need is the Radio Frequency Identification (RFID system. As a result of adopting this technology to the manufacturing industry environment, RFID Network Planning (RNP has become a challenge. Mainly RNP deals with calculating the number and position of antennas which should be deployed in the RFID network to achieve full coverage of the tags that need to be read. The ultimate goal of this paper is to present and evaluate a way of modelling and optimizing nonlinear RNP problems utilizing artificial intelligence (AI techniques. This effort has led the author to propose a novel AI algorithm, which has been named “hybrid AI optimization technique,” to perform optimization of RNP as a hard learning problem. The proposed algorithm is composed of two different optimizationalgorithms: Redundant Antenna Elimination (RAE and Ring Probabilistic Logic Neural Networks (RPLNN. The proposed hybrid paradigm has been explored using a flexible manufacturing system (FMS, and results have been compared with Genetic Algorithm (GA that demonstrates the feasibility of the proposed architecture successfully.

Full Text Available The deployment of passive optical networks (PONs is slow today, especially in Europe, because completely new optical infrastructures are necessary to be installed in the last-mile segments of access networks, which is always very expensive process. One of the possibilities is to design economically effective topologies and to optimize the deployment cost. This article describes the method leading to evaluate an algorithm for designing suboptimal economic solutions and topologies for PONs by focusing on optimization of constructional length of distribution networks. While the typical PON topologies are star topologies or tree-star topologies, the first part of this article introduces new sub algorithm for estimating the minimum star topology. The next section brings the evaluation of two sub algorithms for solving minimum constructional length problems. Finally, all these parts will be merged into a complex algorithm by using clusterization technique to solve optimum topologies. However, the current version of presented algorithm is purely based on mathematical theories and was implemented in Matlab environment. Therefore, it is able to design only theoretical optimum topologies without taking external conditions and real limitations into account. These real conditions will be further implemented in the future, so the algorithm could be also used for practical applications.

Full Text Available This paper proposes the application of a new version of the heuristic particle swarm optimization (PSO method for designing water distribution networks (WDNs. The optimization problem of looped water distribution networks is recognized as an NP-hard combinatorial problem which cannot be easily solved using traditional mathematical optimization techniques. In this paper, the concept of dynamic swarm size is considered in an attempt to increase the convergence speed of the original PSO algorithm. In this strategy, the size of the swarm is dynamically changed according to the iteration number of the algorithm. Furthermore, a novel mutation approach is introduced to increase the diversification property of the PSO and to help the algorithm to avoid trapping in local optima. The new version of the PSO algorithm is called dynamic mutated particle swarm optimization (DMPSO. The proposed DMPSO is then applied to solve WDN design problems. Finally, two illustrative examples are used for comparison to verify the efficiency of the proposed DMPSO as compared to other intelligent algorithms.

Full Text Available Maintaining effective coverage and extending the network lifetime as much as possible has become one of the most critical issues in the coverage of WSNs. In this paper, we propose a multi-objective coverage optimizationalgorithm for WSNs, namely MOCADMA, which models the coverage control of WSNs as the multi-objective optimization problem. MOCADMA uses a memetic algorithm with a dynamic local search strategy to optimize the coverage of WSNs and achieve the objectives such as high network coverage, effective node utilization and more residual energy. In MOCADMA, the alternative solutions are represented as the chromosomes in matrix form, and the optimal solutions are selected through numerous iterations of the evolution process, including selection, crossover, mutation, local enhancement, and fitness evaluation. The experiment and evaluation results show MOCADMA can have good capabilities in maintaining the sensing coverage, achieve higher network coverage while improving the energy efficiency and effectively prolonging the network lifetime, and have a significant improvement over some existing algorithms.

Maintaining effective coverage and extending the network lifetime as much as possible has become one of the most critical issues in the coverage of WSNs. In this paper, we propose a multi-objective coverage optimizationalgorithm for WSNs, namely MOCADMA, which models the coverage control of WSNs as the multi-objective optimization problem. MOCADMA uses a memetic algorithm with a dynamic local search strategy to optimize the coverage of WSNs and achieve the objectives such as high network coverage, effective node utilization and more residual energy. In MOCADMA, the alternative solutions are represented as the chromosomes in matrix form, and the optimal solutions are selected through numerous iterations of the evolution process, including selection, crossover, mutation, local enhancement, and fitness evaluation. The experiment and evaluation results show MOCADMA can have good capabilities in maintaining the sensing coverage, achieve higher network coverage while improving the energy efficiency and effectively prolonging the network lifetime, and have a significant improvement over some existing algorithms. PMID:25360579

Full Text Available The technical and economical evaluation of heat storage layout and configuration in the DH network is one of important aspect for optimizing the heat production from the heat supplier’s point of view in one side as well as to satisfy the heat customer demand in the other side. Generally, the state of the art technique has considered three optional planning layouts for DH network. A classical network with centralized heat storage at Combined Heat and Power (CHP plant, decentralized storages in the network, and decentralized small storages at the substations or in the customer building. In this paper, through the use of genetic algorithm technique, comparison of three different scenarios is presented to evaluate the optimal planning of heat storage layout in CHP based DH supply system according to economical and technical aspects in the network.

The goal of our investigation is optimization of network work in virtual data center. The advantage of modern infrastructure virtualization lies in the possibility to use software-defined networks. However, the existing optimization of algorithmic solutions does not take into account specific features working with multiple classes of virtual network functions. The current paper describes models characterizing the basic structures of object of virtual data center. They including: a level distribution model of software-defined infrastructure virtual data center, a generalized model of a virtual network function, a neural network model of the identification of virtual network functions. We also developed an efficient algorithm for the optimization technology of containerization of virtual network functions in virtual data center. We propose an efficient algorithm for placing virtual network functions. In our investigation we also generalize the well renowned heuristic and deterministic algorithms of Karmakar-Karp.

Full Text Available In this paper, a sensor placement approach to improve the leak location in waterdistribution networks is proposed when the leak signature space (LSS method is used.The sensor placement problem is formulated as an integer optimization problem where thecriterion to be minimized is the number of overlapping signature domains computed fromthe original LSS representation. First, a semi-exhaustive search approach based on a lazyevaluation mechanism ensures optimal placement in the case of low complexity scenarios.For more complex cases, a stochastic optimization process is proposed, based on eitherthe genetic algorithms (GAs or particle swarm optimization (PSO. Experiments on twodifferent networks are used to evaluate the performance of the resolution methods, as well asthe efficiency achieved in the leak location when using the sensor placement results.

Full Text Available Mobile sensor networks are an important part of modern robotics systems and are widely used in robotics applications. Therefore, sensor deployment is a key issue in current robotics systems research. Since it is one of the most popular deployment methods, in recent years the virtual force algorithm has been studied in detail by many scientists. In this paper, we focus on the virtual force algorithm and present a corresponding parameter investigation for mobile sensor deployment. We introduce an optimized virtual force algorithm based on the exchange force, in which a new shielding rule grounded in Delaunay triangulation is adopted. The algorithm employs a new performance metric called ‘pair-correlation diversion', designed to evaluate the uniformity and topology of the sensor distribution. We also discuss the implementation of the algorithm's computation and analyse the influence of experimental parameters on the algorithm. Our results indicate that the area ratio, φs, and the exchange force constant, G, influence the final performance of the sensor deployment in terms of the coverage rate, the convergence time and topology uniformity. Using simulations, we were able to verify the effectiveness of our algorithm and we obtained an optimal region for the (φs, G-parameter space which, in the future, could be utilized as an aid for experiments in robotic sensor deployment.

In this work the results of a genetic algorithm (AG) and a neural recurrent multi state network (RNRME) for optimizing the fuel reload of 5 cycles of the Laguna Verde nuclear power plant (CNLV) are presented. The fuel reload obtained by both methods are compared and it was observed that the RNRME creates better fuel distributions that the AG. Moreover a comparison of the utility for using one or another one techniques is make. (Author)

Large-scale access of distributed power can improve the current environmental pressure, at the same time, increasing the complexity and uncertainty of overall distribution system. Rational planning of distributed power can effectively improve the system voltage level. To this point, the specific impact on distribution network power quality caused by the access of typical distributed power was analyzed and from the point of improving the learning factor and the inertia weight, an improved particle swarm optimizationalgorithm (IPSO) was proposed which could solve distributed generation planning for distribution network to improve the local and global search performance of the algorithm. Results show that the proposed method can well reduce the system network loss and improve the economic performance of system operation with distributed generation.

Energy efficiency aspects in cellular networks can contribute significantly to reducing worldwide greenhouse gas emissions. The base station (BS) sleeping strategy has become a well-known technique to achieve energy savings by switching off redundant BSs mainly for lightly loaded networks. Moreover, introducing renewable energy as an alternative power source has become a real challenge among network operators. In this paper, we formulate an optimization problem that aims to maximize the profit of Long-Term Evolution (LTE) cellular operators and to simultaneously minimize the CO2 emissions in green wireless cellular networks without affecting the desired quality of service (QoS). The BS sleeping strategy lends itself to an interesting implementation using several heuristic approaches, such as the genetic (GA) and particle swarm optimization (PSO) algorithms. In this paper, we propose GA-based and PSO-based methods that reduce the energy consumption of BSs by not only shutting down underutilized BSs but by optimizing the amounts of energy procured from different retailers (renewable energy and electricity retailers), as well. A comparison with another previously proposed algorithm is also carried out to evaluate the performance and the computational complexity of the employed methods.

For meeting the real-time fault diagnosis and the optimization monitoring requirements of the polymerization kettle in the polyvinyl chloride resin (PVC) production process, a fault diagnosis strategy based on the self-organizing map (SOM) neural network is proposed. Firstly, a mapping between the polymerization process data and the fault pattern is established by analyzing the production technology of polymerization kettle equipment. The particle swarm optimization (PSO) algorithm with a new dynamical adjustment method of inertial weights is adopted to optimize the structural parameters of SOM neural network. The fault pattern classification of the polymerization kettle equipment is to realize the nonlinear mapping from symptom set to fault set according to the given symptom set. Finally, the simulation experiments of fault diagnosis are conducted by combining with the industrial on-site historical data of the polymerization kettle and the simulation results show that the proposed PSO-SOM fault diagnosis strategy is effective.

Full Text Available For meeting the real-time fault diagnosis and the optimization monitoring requirements of the polymerization kettle in the polyvinyl chloride resin (PVC production process, a fault diagnosis strategy based on the self-organizing map (SOM neural network is proposed. Firstly, a mapping between the polymerization process data and the fault pattern is established by analyzing the production technology of polymerization kettle equipment. The particle swarm optimization (PSO algorithm with a new dynamical adjustment method of inertial weights is adopted to optimize the structural parameters of SOM neural network. The fault pattern classification of the polymerization kettle equipment is to realize the nonlinear mapping from symptom set to fault set according to the given symptom set. Finally, the simulation experiments of fault diagnosis are conducted by combining with the industrial on-site historical data of the polymerization kettle and the simulation results show that the proposed PSO-SOM fault diagnosis strategy is effective.

A district cooling system is a sustainable means of distribution of cooling energy through mass production. A cooling medium like chilled water is generated at a central refrigeration plant and supplied to serve a group of consumer buildings through a piping network. Because of the substantial capital investment involved, an optimal design of the distribution piping configuration is one of the crucial factors for successful implementation of the district cooling scheme. In the present study, genetic algorithm (GA) incorporated with local search techniques was developed to find the optimal/near optimal configuration of the piping network in a hypothetical site. The effect of local search, mutation rate and frequency of local search on the performance of the GA in terms of both solution quality and computation time were investigated and presented in this paper

Full Text Available Seyedali Mirjalili et al. (2014 introduced a completely unique metaheuristic technique particularly grey wolf optimization (GWO. This algorithm mimics the social behavior of grey wolves whereas it follows the leadership hierarchy and attacking strategy. The rising issue in wireless sensor network (WSN is localization problem. The objective of this problem is to search out the geographical position of unknown nodes with the help of anchor nodes in WSN. In this work, GWO algorithm is incorporated to spot the correct position of unknown nodes, so as to handle the node localization problem. The proposed work is implemented using MATLAB 8.2 whereas nodes are deployed in a random location within the desired network area. The parameters like computation time, percentage of localized node, and minimum localization error measures are utilized to analyse the potency of GWO rule with other variants of metaheuristics algorithms such as particle swarm optimization (PSO and modified bat algorithm (MBA. The observed results convey that the GWO provides promising results compared to the PSO and MBA in terms of the quick convergence rate and success rate.

Full Text Available Differential evolution (DE algorithm is used to determine optimal location of unified power quality conditioner (UPQC considering its size in the radial distribution systems. The problem is formulated to find the optimum location of UPQC based on an objective function (OF defined for improving of voltage and current profiles, reducing power loss and minimizing the investment costs considering the OF's weighting factors. Hence, a steady-state model of UPQC is derived to set in forward/backward sweep load flow. Studies are performed on two IEEE 33-bus and 69-bus standard distribution networks. Accuracy was evaluated by reapplying the procedures using both genetic (GA and immune algorithms (IA. Comparative results indicate that DE is capable of offering a nearer global optimal in minimizing the OF and reaching all the desired conditions than GA and IA.

Full Text Available Najla S Dar-Odeh1, Othman M Alsmadi2, Faris Bakri3, Zaer Abu-Hammour2, Asem A Shehabi3, Mahmoud K Al-Omiri1, Shatha M K Abu-Hammad4, Hamzeh Al-Mashni4, Mohammad B Saeed4, Wael Muqbil4, Osama A Abu-Hammad1 1Faculty of Dentistry, 2Faculty of Engineering and Technology, 3Faculty of Medicine, University of Jordan, Amman, Jordan; 4Dental Department, University of Jordan Hospital, Amman, JordanObjective: To construct and optimize a neural network that is capable of predicting the occurrence of recurrent aphthous ulceration (RAU based on a set of appropriate input data.Participants and methods: Artificial neural networks (ANN software employing genetic algorithms to optimize the architecture neural networks was used. Input and output data of 86 participants (predisposing factors and status of the participants with regards to recurrent aphthous ulceration were used to construct and train the neural networks. The optimized neural networks were then tested using untrained data of a further 10 participants.Results: The optimized neural network, which produced the most accurate predictions for the presence or absence of recurrent aphthous ulceration was found to employ: gender, hematological (with or without ferritin and mycological data of the participants, frequency of tooth brushing, and consumption of vegetables and fruits.Conclusions: Factors appearing to be related to recurrent aphthous ulceration and appropriate for use as input data to construct ANNs that predict recurrent aphthous ulceration were found to include the following: gender, hemoglobin, serum vitamin B12, serum ferritin, red cell folate, salivary candidal colony count, frequency of tooth brushing, and the number of fruits or vegetables consumed daily.Keywords: artifical neural networks, recurrent, aphthous ulceration, ulcer

Wind tunnels use scale models to characterize aerodynamic coefficients, Wind tunnel testing can be slow and costly due to high personnel overhead and intensive power utilization. Although manual curve fitting can be done, it is highly efficient to use a neural network to define the complex relationship between variables. Numerical simulation of complex vehicles on the wide range of conditions required for flight simulation requires static and dynamic data. Static data at low Mach numbers and angles of attack may be obtained with simpler Euler codes. Static data of stalled vehicles where zones of flow separation are usually present at higher angles of attack require Navier-Stokes simulations which are costly due to the large processing time required to attain convergence. Preliminary dynamic data may be obtained with simpler methods based on correlations and vortex methods; however, accurate prediction of the dynamic coefficients requires complex and costly numerical simulations. A reliable and fast method of predicting complex aerodynamic coefficients for flight simulation I'S presented using a neural network. The training data for the neural network are derived from numerical simulations and wind-tunnel experiments. The aerodynamic coefficients are modeled as functions of the flow characteristics and the control surfaces of the vehicle. The basic coefficients of lift, drag and pitching moment are expressed as functions of angles of attack and Mach number. The modeled and training aerodynamic coefficients show good agreement. This method shows excellent potential for rapid development of aerodynamic models for flight simulation. Genetic Algorithms (GA) are used to optimize a previously built Artificial Neural Network (ANN) that reliably predicts aerodynamic coefficients. Results indicate that the GA provided an efficient method of optimizing the ANN model to predict aerodynamic coefficients. The reliability of the ANN using the GA includes prediction of aerodynamic

Full Text Available The trial-and-error method is widely used for the current optimization of the steel casting feeding system, which is highly random, subjective and thus inefficient. In the present work, both the theoretical and the experimental research on the modeling and optimization methods of the process are studied. An approximate alternative model is established based on the Back Propagation (BP neural network and experimental design. The process parameters of the feeding system are taken as the input, the volumes of shrinkage cavities and porosities calculated by simulation are simultaneously taken as the output. Thus, a mathematical model is established by the BP neural network to combine the input variables with the output response. Then, this model is optimized by the nonlinear optimization function of the genetic algorithm. Finally, a feeding system optimization of a steel traveling wheel is conducted. No shrinkage cavities and porosities are induced through the optimization. Compared to the initial design scheme, the process yield is increased by 4.1% and the volume of the riser is decreased by 5.48×106 mm3.

A back-propagation (BP) neural network can solve complicated random nonlinear mapping problems; therefore, it can be applied to a wide range of problems. However, as the sample size increases, the time required to train BP neural networks becomes lengthy. Moreover, the classification accuracy decreases as well. To improve the classification accuracy and runtime efficiency of the BP neural networkalgorithm, we proposed a parallel design and realization method for a particle swarm optimization (PSO)-optimized BP neural network based on MapReduce on the Hadoop platform using both the PSO algorithm and a parallel design. The PSO algorithm was used to optimize the BP neural network's initial weights and thresholds and improve the accuracy of the classification algorithm. The MapReduce parallel programming model was utilized to achieve parallel processing of the BP algorithm, thereby solving the problems of hardware and communication overhead when the BP neural network addresses big data. Datasets on 5 different scales were constructed using the scene image library from the SUN Database. The classification accuracy of the parallel PSO-BP neural networkalgorithm is approximately 92%, and the system efficiency is approximately 0.85, which presents obvious advantages when processing big data. The algorithm proposed in this study demonstrated both higher classification accuracy and improved time efficiency, which represents a significant improvement obtained from applying parallel processing to an intelligent algorithm on big data.

The quantitative elemental content analysis is difficult due to the uniform effect, particle effect and the element matrix effect, etc, when using energy dispersive X-ray fluorescence (EDXRF) technique. In this paper, a hybrid approach of genetic algorithm (GA) and back propagation (BP) neural network was proposed without considering the complex relationship between the concentration and intensity. The aim of GA optimized BP was to get better network initial weights and thresholds. The basic idea was that the reciprocal of the mean square error of the initialization BP neural network was set as the fitness value of the individual in GA, and the initial weights and thresholds were replaced by individuals, and then the optimal individual was sought by selection, crossover and mutation operations, finally a new BP neural network model was created with the optimal initial weights and thresholds. The calculation results of quantitative analysis of titanium and iron contents for five types of ore bodies in Panzhihua Mine show that the results of classification prediction are far better than that of overall forecasting, and relative errors of 76.7% samples are less than 2% compared with chemical analysis values, which demonstrates the effectiveness of the proposed method. (authors)

Natural Gas is a major energy source in Europe, yet political instabilities have the potential to disrupt access and supply. Energy resilience is an increasingly essential construct and begins with transmission network design. This study proposes a new way of thinking about modelling natural gas flow. Rather than relying on classical economic models, this problem is cast into a time-dependent Hamiltonian dynamics discussion. Traditional Natural Gas constraints, including inelastic demand and maximum/minimum pipe flows, are portrayed as energy functions and built into the dynamics of each pipe flow. Doing so allows the constraints to be built into the dynamics of each pipeline. As time progresses in the model, natural gas flow rates find the minimum energy, thus the optimal gas flow rates. The most important result of this study is using dynamical principles to ensure the output of natural gas at demand nodes remains constant, which is important for country to country natural gas transmission. Another important step in this study is building the dynamics of each flow in a decentralized algorithm format. Decentralized regulation has solved congestion problems for internet data flow, traffic flow, epidemiology, and as demonstrated in this study can solve the problem of Natural Gas congestion. A mathematical description is provided for how decentralized regulation leads to globally optimizednetwork flow. Furthermore, the dynamical principles and decentralized algorithm are applied to a case study of the Fluxys Belgium Natural Gas Network.

Full Text Available This article reflects algorithmization of search methods of effective replacement of consumer transformers in distributed electrical networks. As any electrical equipment of power systems, power transformers have their own limited service duration, which is determined by natural processes of materials degradation and also by unexpected wear under different conditions of overload and overvoltage. According to the standards, adapted by in the Republic of Belarus, rated service life of power transformers is 25 years. But it can be situations that transformers should be better changed till this time – economically efficient. The possibility of such replacement is considered in order to increase efficiency of electrical network operation connected with its physical wear and aging.In this article the faults of early developed mathematical models of transformers replacement were discussed. Early such worked out transformers were not used. But in practice they can be replaced in one substation but they can be successfully used in other substations .Especially if there are limits of financial resources and the replacement needs more detail technical and economical basis.During the research the authors developed the efficient algorithm for determining of optimal location of transformers at substations of distributed electrical networks, based on search of the best solution from all sets of displacement in oriented graph. Suggested algorithm allows considerably reduce design time of optimal placement of transformers using a set of simplifications. The result of algorithm’s work is series displacement of transformers in networks, which allow obtain a great economic effect in comparison with replacement of single transformer.

Full Text Available In recent years, heterogeneous radio access technologies have experienced rapid development and gradually achieved effective coordination and integration, resulting in heterogeneous networks (HetNets. In this paper, we consider the downlink secure transmission of HetNets where the information transmission from base stations (BSs to legitimate users is subject to the interception of eavesdroppers. In particular, we stress the problem of joint user association and power allocation of the BSs. To achieve data transmission in a secure and energy efficient manner, we introduce the concept of secrecy energy efficiency which is defined as the ratio of the secrecy transmission rate and power consumption of the BSs and formulate the problem of joint user association and power allocation as an optimization problem which maximizes the joint secrecy energy efficiency of all the BSs under the power constraint of the BSs and the minimum data rate constraint of user equipment (UE. By equivalently transforming the optimization problem into two subproblems, that is, power allocation subproblem and user association subproblem of the BSs, and applying iterative method and Kuhn-Munkres (K-M algorithm to solve the two subproblems, respectively, the optimal user association and power allocation strategies can be obtained. Numerical results demonstrate that the proposed algorithm outperforms previously proposed algorithms.

Full Text Available Stock prediction is to determine the future value of a company stock dealt on an exchange. It plays a crucial role to raise the profit gained by firms and investors. Over the past few years, many methods have been developed in which plenty of efforts focus on the machine learning framework achieving the promising results. In this paper, an approach based on Artificial Neural Network (ANN optimized by Fireworks algorithm and data preprocessing by Haar Wavelet is applied to estimate the stock prices. The system was trained and tested with real data of various companies collected from Yahoo Finance. The obtained results are encouraging.

Full Text Available Regional logistics prediction is the key step in regional logistics planning and logistics resources rationalization. Since regional economy is the inherent and determinative factor of regional logistics demand, it is feasible to forecast regional logistics demand by investigating economic indicators which can accelerate the harmonious development of regional logistics industry and regional economy. In this paper, the PSO-RBFNN model, a radial basis function neural network (RBFNN combined with particle swarm optimization (PSO algorithm, is studied. The PSO-RBFNN model is trained by indicators data in a region to predict the regional logistics demand. And the corresponding results indicate the model’s applicability and potential advantages.

Artificial Intelligence is a big part of automation and with today's technological advances, artificial intelligence has taken great strides towards positioning itself as the technology of the future to control, enhance and perfect automation. Computer vision includes pattern recognition and classification and machine learning. Computer vision is at the core of decision making and it is a vast and fruitful branch of artificial intelligence. In this work, we expose novel algorithms and techniques built upon existing technologies to improve pattern recognition and neural network training, initially motivated by a multidisciplinary effort to build a robot that helps maintain and optimize solar panel energy production. Our contributions detail an improved non-linear pre-processing technique to enhance poorly illuminated images based on modifications to the standard histogram equalization for an image. While the original motivation was to improve nocturnal navigation, the results have applications in surveillance, search and rescue, medical imaging enhancing, and many others. We created a vision system for precise camera distance positioning motivated to correctly locate the robot for capture of solar panel images for classification. The classification algorithm marks solar panels as clean or dirty for later processing. Our algorithm extends past image classification and, based on historical and experimental data, it identifies the optimal moment in which to perform maintenance on marked solar panels as to minimize the energy and profit loss. In order to improve upon the classification algorithm, we delved into feedforward neural networks because of their recent advancements, proven universal approximation and classification capabilities, and excellent recognition rates. We explore state-of-the-art neural network training techniques offering pointers and insights, culminating on the implementation of a complete library with support for modern deep learning architectures

Full Text Available The goal of this study is to improve thermal comfort and indoor air quality with the adaptive network-based fuzzy inference system (ANFIS model and improved particle swarm optimization (PSO algorithm. A method to optimize air conditioning parameters and installation distance is proposed. The methodology is demonstrated through a prototype case, which corresponds to a typical laboratory in colleges and universities. A laboratory model is established, and simulated flow field information is obtained with the CFD software. Subsequently, the ANFIS model is employed instead of the CFD model to predict indoor flow parameters, and the CFD database is utilized to train ANN input-output “metamodels” for the subsequent optimization. With the improved PSO algorithm and the stratified sequence method, the objective functions are optimized. The functions comprise PMV, PPD, and mean age of air. The optimal installation distance is determined with the hemisphere model. Results show that most of the staff obtain a satisfactory degree of thermal comfort and that the proposed method can significantly reduce the cost of building an experimental device. The proposed methodology can be used to determine appropriate air supply parameters and air conditioner installation position for a pleasant and healthy indoor environment.

Artificial neural networks (Ann) are widely used; it which consist of an input layer, one or more hidden layers and an output layer; these layers contain neurons and each has connections called weights, where the knowledge are allowed and let to Ann solve problems proposed. These Ann is used to reconstruction of the energy spectrum of neutrons from count rates and develop Bonner sphere neutron dosimetry. Currently, we have developed Ann with high performance and generalization ability. Determine your optimal architecture is usually a difficult task, an exhaustive search of all possible combinations of parameters is rarely possible further training of the neural network with random initial weights can cause two major drawbacks: it can stuck in local minima or converge very slowly. In this project it will be used Genetic Algorithms (Ga); which are based on the principle or analogy of evolution through natural selection and has shown to be very effective in optimizing complex search functions and large spaces or to find a near optimal overall efficiency. The aim is to decrease the architecture in number of hidden neurons and therefore the total number of connections is reducing. The benefits obtained by optimizing the network are that the number of connections would be considerably smaller and thus the computational complexity, hardware integration, resources will be lower such that will allow to be even more viable implemented. To use the Ga three problems must be solve: 1) coding the problem into chromosomes. 2) Construct a fitness function. 3) Proper selection of genetic operators; crossover, selection, mutation. As a result, the scientific knowledge obtained can to be applied to similar problems having a reference parameters used and their impact on the optimization would to be generated. It concluded that the input layer and output are subject to the problem; the Ga propose the optimal number of neurons in the hidden layer without losing the quality of the

Full Text Available This paper proposes a new sensor placement approach for leak location in water distribution networks (WDNs. The sensor placement problem is formulated as an integer optimization problem. The optimization criterion consists in minimizing the number of non-isolable leaks according to the isolability criteria introduced. Because of the large size and non-linear integer nature of the resulting optimization problem, genetic algorithms (GAs are used as the solution approach. The obtained results are compared with a semi-exhaustive search method with higher computational effort, proving that GA allows one to find near-optimal solutions with less computational load. Moreover, three ways of increasing the robustness of the GA-based sensor placement method have been proposed using a time horizon analysis, a distance-based scoring and considering different leaks sizes. A great advantage of the proposed methodology is that it does not depend on the isolation method chosen by the user, as long as it is based on leak sensitivity analysis. Experiments in two networks allow us to evaluate the performance of the proposed approach.

This paper proposes a new sensor placement approach for leak location in water distribution networks (WDNs). The sensor placement problem is formulated as an integer optimization problem. The optimization criterion consists in minimizing the number of non-isolable leaks according to the isolability criteria introduced. Because of the large size and non-linear integer nature of the resulting optimization problem, genetic algorithms (GAs) are used as the solution approach. The obtained results are compared with a semi-exhaustive search method with higher computational effort, proving that GA allows one to find near-optimal solutions with less computational load. Moreover, three ways of increasing the robustness of the GA-based sensor placement method have been proposed using a time horizon analysis, a distance-based scoring and considering different leaks sizes. A great advantage of the proposed methodology is that it does not depend on the isolation method chosen by the user, as long as it is based on leak sensitivity analysis. Experiments in two networks allow us to evaluate the performance of the proposed approach.

Full Text Available Power system network can undergo outages during which there may be a partial or total blackout in the system. In that condition, transmission of power through the optimal path is an important problem in the process of reconfiguration of power system components. For a given set of generation, load pair, there could be many possible paths to transmit the power. The optimal path needs to consider the shortest path (minimum losses, capacity of the transmission line, voltage stability, priority of loads, and power balance between the generation and demand. In this paper, the Bellman Ford Algorithm (BFA is applied to find out the optimal path and also the several alternative paths by considering all the constraints. In order to demonstrate the capability of BFA, it has been applied to a practical 230 kV network. This restorative path search guidance tool is quite efficient in finding the optimal and also the alternate paths for transmitting the power from a generating station to demand.

An innovative computational approach for the optimization of groundwater remediation is presented which uses artificial neural networks (ANNs) and the genetic algorithm (GA). In this approach, the ANN is trained to predict an aspect of the outcome of a flow and transport simulation. Then the GA searches through realizations or patterns of pumping and uses the trained network to predict the outcome of the realizations. This approach has advantages of parallel processing of the groundwater simulations and the ability to ''recycle'' or reuse the base of knowledge formed by these simulations. These advantages offer reduction of computational burden of the groundwater simulations relative to a more conventional approach which uses nonlinear programming (NLP) with a quasi-newtonian search. Also the modular nature of this approach facilitates substitution of different groundwater simulation models

An innovative computational approach for the optimization of groundwater remediation is presented which uses artificial neural networks (ANNs) and the genetic algorithm (GA). In this approach, the ANN is trained to predict an aspect of the outcome of a flow and transport simulation. Then the GA searches through realizations or patterns of pumping and uses the trained network to predict the outcome of the realizations. This approach has advantages of parallel processing of the groundwater simulations and the ability to ``recycle`` or reuse the base of knowledge formed by these simulations. These advantages offer reduction of computational burden of the groundwater simulations relative to a more conventional approach which uses nonlinear programming (NLP) with a quasi-newtonian search. Also the modular nature of this approach facilitates substitution of different groundwater simulation models.

Supercritical CO 2 power cycle shows a high potential to recover low-grade waste heat due to its better temperature glide matching between heat source and working fluid in the heat recovery vapor generator (HRVG). Parametric analysis and exergy analysis are conducted to examine the effects of thermodynamic parameters on the cycle performance and exergy destruction in each component. The thermodynamic parameters of the supercritical CO 2 power cycle is optimized with exergy efficiency as an objective function by means of genetic algorithm (GA) under the given waste heat condition. An artificial neural network (ANN) with the multi-layer feed-forward network type and back-propagation training is used to achieve parametric optimization design rapidly. It is shown that the key thermodynamic parameters, such as turbine inlet pressure, turbine inlet temperature and environment temperature have significant effects on the performance of the supercritical CO 2 power cycle and exergy destruction in each component. It is also shown that the optimum thermodynamic parameters of supercritical CO 2 power cycle can be predicted with good accuracy using artificial neural network under variable waste heat conditions.

The metabolic network model allows for an in-depth insight into the molecular mechanism of a particular organism. Because most parameters of the metabolic network cannot be directly measured, they must be estimated by using optimizationalgorithms. However, three characteristics of the metabolic network model, i.e., high nonlinearity, large amount parameters, and huge variation scopes of parameters, restrict the application of many traditional optimizationalgorithms. As a result, there is a growing demand to develop efficient optimization approaches to address this complex problem. In this paper, a Kriging-based algorithm aiming at parameter estimation is presented for constructing the metabolic networks. In the algorithm, a new infill sampling criterion, named expected improvement and mutual information (EI&MI), is adopted to improve the modeling accuracy by selecting multiple new sample points at each cycle, and the domain decomposition strategy based on the principal component analysis is introduced to save computing time. Meanwhile, the convergence speed is accelerated by combining a single-dimensional optimization method with the dynamic coordinate perturbation strategy when determining the new sample points. Finally, the algorithm is applied to the arachidonic acid metabolic network to estimate its parameters. The obtained results demonstrate the effectiveness of the proposed algorithm in getting precise parameter values under a limited number of iterations.

Feature extraction and classification of electrocardiogram (ECG) signals are necessary for the automatic diagnosis of cardiac diseases. In this study, a novel method based on genetic algorithm-back propagation neural network (GA-BPNN) for classifying ECG signals with feature extraction using wavelet packet decomposition (WPD) is proposed. WPD combined with the statistical method is utilized to extract the effective features of ECG signals. The statistical features of the wavelet packet coefficients are calculated as the feature sets. GA is employed to decrease the dimensions of the feature sets and to optimize the weights and biases of the back propagation neural network (BPNN). Thereafter, the optimized BPNN classifier is applied to classify six types of ECG signals. In addition, an experimental platform is constructed for ECG signal acquisition to supply the ECG data for verifying the effectiveness of the proposed method. The GA-BPNN method with the MIT-BIH arrhythmia database achieved a dimension reduction of nearly 50% and produced good classification results with an accuracy of 97.78%. The experimental results based on the established acquisition platform indicated that the GA-BPNN method achieved a high classification accuracy of 99.33% and could be efficiently applied in the automatic identification of cardiac arrhythmias.

Full Text Available In a nuclear industry area, high temperature treatment of materials is a factor which requires special attention. Assessment needs to be conducted on the properties of the materials used, including the strength of the materials. The measurement of material properties under thermal processes may reflect residual stresses. The use of Genetic Algorithm (GA to determine the optimal residual stress is one way to determine the strength of a material. In residual stress modeling with several parameters, it is sometimes difficult to solve for the optimal value through analytical or numerical calculations. Here, GA is an efficient algorithm which can generate the optimal values, both minima and maxima. The purposes of this research are to obtain the optimization of variable in residual stress models using GA and to predict the center of residual stress distribution, using fuzzy neural network (FNN while the artificial neural network (ANN used for modeling. In this work a single-material 316/316L stainless steel bar is modeled. The minimal residual stresses of the material at high temperatures were obtained with GA and analytical calculations. At a temperature of 6500C, the GA optimal residual stress estimation converged at –711.3689 MPa at adistance of 0.002934 mm from center point, whereas the analytical calculation result at that temperature and position is -975.556 MPa . At a temperature of 8500C, the GA result was -969.868 MPa at 0.002757 mm from the center point, while with analytical result was -1061.13 MPa. The difference in residual stress between GA and analytical results at a temperatureof6500C is about 27 %, while at 8500C it is 8.67 %. The distribution of residual stress showed a grouping concentrated around a coordinate of (-76; 76 MPa. The residuals stress model is a degree-two polynomial with coefficients of 50.33, -76.54, and -55.2, respectively, with a standard deviation of 7.874.

Prepared ca. 1995. This paper illustrates the use of the simulation-optimization technique of response surface methodology (RSM) in traffic signal optimization of urban networks. It also quantifies the gains of using the common random number (CRN) va...

Full Text Available The fault-tolerant routing problem is important consideration in the design of heterogeneous wireless sensor networks (H-WSNs applications, and has recently been attracting growing research interests. In order to maintain k disjoint communication paths from source sensors to the macronodes, we present a hybrid routing scheme and model, in which multiple paths are calculated and maintained in advance, and alternate paths are created once the previous routing is broken. Then, we propose an immune cooperative particle swarm optimizationalgorithm (ICPSOA in the model to provide the fast routing recovery and reconstruct the network topology for path failure in H-WSNs. In the ICPSOA, mutation direction of the particle is determined by multi-swarm evolution equation, and its diversity is improved by immune mechanism, which can enhance the capacity of global search and improve the converging rate of the algorithm. Then we validate this theoretical model with simulation results. The results indicate that the ICPSOA-based fault-tolerant routing protocol outperforms several other protocols due to its capability of fast routing recovery mechanism, reliable communications, and prolonging the lifetime of WSNs.

Full Text Available This paper presents the model and algorithms for traffic flow data monitoring and optimal traffic light control based on wireless sensor networks. Given the scenario that sensor nodes are sparsely deployed along the segments between signalized intersections, an analytical model is built using continuum traffic equation and develops the method to estimate traffic parameter with the scattered sensor data. Based on the traffic data and principle of traffic congestion formation, we introduce the congestion factor which can be used to evaluate the real-time traffic congestion status along the segment and to predict the subcritical state of traffic jams. The result is expected to support the timing phase optimization of traffic light control for the purpose of avoiding traffic congestion before its formation. We simulate the traffic monitoring based on the Mobile Century dataset and analyze the performance of traffic light control on VISSIM platform when congestion factor is introduced into the signal timing optimization model. The simulation result shows that this method can improve the spatial-temporal resolution of traffic data monitoring and evaluate traffic congestion status with high precision. It is helpful to remarkably alleviate urban traffic congestion and decrease the average traffic delays and maximum queue length.

Recently, the possibility of occurrence of localized strong heavy rainfall due to climate change is increasing and flood damage is also increasing trend in Korea. Therefore we need more precise hydrologic analysis for preparing alternatives or measures for flood reduction by considering climate conditions which we have difficulty in the prediction. To do this, obtaining reliable hydrologic data, for an example, stage data, is very important. However, the existing stage gauge stations are scattered around the country, making it difficult to maintain them in a stable manner, and subsequently hard to acquire the hydrologic data that could be used for reflecting the localized hydrologic characteristics. In order to overcome such restrictions, this paper not only aims to establish a plan to acquire the water stage data in a constant and proper manner by using limited manpower and costs, but also establishes the fundamental technology for acquiring the water level observation data or the stage data. For that, this paper identifies the current status of the stage gauge stations installed in the Chung-Ju dam in Han river, Korea and extract the factors related to the division and characteristics of basins. Then, the obtained factors are used to develop the representative unit hydrograph that shows the characteristics of flow. After that, the data are converted into the probability density function and the stations at individual basins are selected by using the entropy theory. In last step, we establish the optimized stage gauge network by the location of the stage station and grade using the Multi Objective Genetic Algorithm(MOGA) technique that takes into account for the combinations of the number of the stations. It is expected that this paper can help establish an optimal observational network of stage guages as it can be applied usefully not only for protecting against floods in a stable manner, but also for acquiring the hydrologic data in an efficient manner. Keywords

algorithm using successive linear programming is presented. The performance of the algorithm is illustrated by optimizing a network with 201 pipes and 172 nodes. It is concluded that the new algorithm seems to be very efficient and stable, and that it always finds a solution with a cost near the best...

The computational evolution of gene networks functions like a forward genetic screen to generate, without preconceptions, all networks that can be assembled from a defined list of parts to implement a given function. Frequently networks are subject to multiple design criteria that cannot all be optimized simultaneously. To explore how these tradeoffs interact with evolution, we implement Pareto optimization in the context of gene network evolution. In response to a temporal pulse of a signal, we evolve networks whose output turns on slowly after the pulse begins, and shuts down rapidly when the pulse terminates. The best performing networks under our conditions do not fall into categories such as feed forward and negative feedback that also encode the input–output relation we used for selection. Pareto evolution can more efficiently search the space of networks than optimization based on a single ad hoc combination of the design criteria. (paper)

The computational evolution of gene networks functions like a forward genetic screen to generate, without preconceptions, all networks that can be assembled from a defined list of parts to implement a given function. Frequently networks are subject to multiple design criteria that cannot all be optimized simultaneously. To explore how these tradeoffs interact with evolution, we implement Pareto optimization in the context of gene network evolution. In response to a temporal pulse of a signal, we evolve networks whose output turns on slowly after the pulse begins, and shuts down rapidly when the pulse terminates. The best performing networks under our conditions do not fall into categories such as feed forward and negative feedback that also encode the input-output relation we used for selection. Pareto evolution can more efficiently search the space of networks than optimization based on a single ad hoc combination of the design criteria.

Full Text Available Effective ventilation planning is vital to underground mining. To ensure stable operation of the ventilation system and to avoid airflow disorder, mine ventilation network (MVN models have been widely used in simulating and optimizing the mine ventilation system. However, one of the challenges for MVN model simulation is that the simulated airflow distribution results do not match the measured data. To solve this problem, a simple and effective calibration method is proposed based on the non-linear optimizationalgorithm. The calibrated model not only makes simulated airflow distribution results in accordance with the on-site measured data, but also controls the errors of other parameters within a minimum range. The proposed method was then applied to calibrate an MVN model in a real case, which is built based on ventilation survey results and Ventsim software. Finally, airflow simulation experiments are carried out respectively using data before and after calibration, whose results were compared and analyzed. This showed that the simulated airflows in the calibrated model agreed much better to the ventilation survey data, which verifies the effectiveness of calibrating method.

The use of evolutionary algorithms for global optimization has increased rapidly during the past several years. But evolutionary computations have a common drawback: they need a huge number of function evaluations. This makes them inadequate for structural optimization. To overcome this difficulty, the authors propose a method that integrates the evolutionary algorithm with data mining and approximate analysis to find the optimal solution in structural optimization. The approximate analysis is used to replace exact finite element analyses and the data mining is employed to identify feasible solutions. These combined efforts can reduce the computational time and search the feasible region intensively. As a result, the efficiency and quality of structural optimization using evolutionary algorithms will be increased. Some test problems show that the proposed method not only finds the global solution but is also less computationally demanding.

Full Text Available With the development of wireless technologies, mobile communication applies more and more extensively in the various walks of life. The social network of both fixed and mobile users can be seen as networked agent system. At present, kinds of devices and access network technology are widely used. Different users in this networked agent system may need different coding rates multimedia data due to their heterogeneous demand. This paper proposes a distributed flow rate control algorithm to optimize multimedia data transmission of the networked agent system with the coexisting various coding rates. In this proposed algorithm, transmission path and upload bandwidth of different coding rate data between source node, fixed and mobile nodes are appropriately arranged and controlled. On the one hand, this algorithm can provide user nodes with differentiated coding rate data and corresponding flow rate. On the other hand, it makes the different coding rate data and user nodes networked, which realizes the sharing of upload bandwidth of user nodes which require different coding rate data. The study conducts mathematical modeling on the proposed algorithm and compares the system that adopts the proposed algorithm with the existing system based on the simulation experiment and mathematical analysis. The results show that the system that adopts the proposed algorithm achieves higher upload bandwidth utilization of user nodes and lower upload bandwidth consumption of source node.

To improve the traditional smart home system, its electric wiring, networking technology, information transmission and facility control are studied. In this paper, we study the electric wiring, networking technology, information transmission and facility control to improve the traditional smart home system. First, ZigBee is used to replace the traditional electric wiring. Second, a network is built to connect lots of wireless sensors and facilities, thanks to the capability of ZigBee self-organized network and Genetic Algorithm-Particle Swarm OptimizationAlgorithm (GA-PSOA) to search for the optimal route. Finally, when the smart home system is connected to the internet based on the remote server technology, home environment and facilities could be remote real-time controlled. The experiments show that the GA-PSOA reduce the system delay and decrease the energy consumption of the wireless system.

Full Text Available This paper proposes a Firefly algorithm based technique to optimize the control variables for simultaneous optimization of real power loss and voltage stability limit of the transmission system. Mathematically, this issue can be formulated as nonlinear equality and inequality constrained optimization problem with an objective function integrating both real power loss and voltage stability limit. Transformers taps, unified power flow controller and its parameters have been included as control variables in the problem formulation. The effectiveness of the proposed algorithm has been tested on New England 39-bus system. Simulation results obtained with the proposed algorithm are compared with the real coded genetic algorithm for single objective of real power loss minimization and multi-objective of real power loss minimization and voltage stability limit maximization. Also, a classical optimization method known as interior point successive linear programming technique is considered here to compare the results of firefly algorithm for single objective of real power loss minimization. Simulation results confirm the potentiality of the proposed algorithm in solving optimization problems.

Full Text Available Data collection is a fundamental operation in various mobile wireless sensor networks (MWSN applications. The energy of nodes around the Sink can be untimely depleted owing to the fact that sensor nodes must transmit vast amounts of data, readily forming a bottleneck in energy consumption; mobile wireless sensor networks have been designed to address this issue. In this study, we focused on a large-scale and intensive MWSN which allows a certain amount of data latency by investigating mobile Sink balance from three aspects: data collection maximization, mobile path length minimization, and network reliability optimization. We also derived a corresponding formula to represent the MWSN and proved that it represents an NP-hard problem. Traditional data collection methods only focus on increasing the amount data collection or reducing the overall network energy consumption, which is why we designed the proposed heuristic algorithm to jointly consider cluster head selection, the routing path from ordinary nodes to the cluster head node, and mobile Sink path planning optimization. The proposed data collection algorithm for mobile Sinks is, in effect, based on artificial bee colony. Simulation results show that, in comparison with other algorithms, the proposed algorithm can effectively reduce data transmission, save energy, improve network data collection efficiency and reliability, and extend the network lifetime.

Full Text Available Abstract Background Reverse engineering gene networks and identifying regulatory interactions are integral to understanding cellular decision making processes. Advancement in high throughput experimental techniques has initiated innovative data driven analysis of gene regulatory networks. However, inherent noise associated with biological systems requires numerous experimental replicates for reliable conclusions. Furthermore, evidence of robust algorithms directly exploiting basic biological traits are few. Such algorithms are expected to be efficient in their performance and robust in their prediction. Results We have developed a network identification algorithm to accurately infer both the topology and strength of regulatory interactions from time series gene expression data in the presence of significant experimental noise and non-linear behavior. In this novel formulism, we have addressed data variability in biological systems by integrating network identification with the bootstrap resampling technique, hence predicting robust interactions from limited experimental replicates subjected to noise. Furthermore, we have incorporated non-linearity in gene dynamics using the S-system formulation. The basic network identification formulation exploits the trait of sparsity of biological interactions. Towards that, the identification algorithm is formulated as an integer-programming problem by introducing binary variables for each network component. The objective function is targeted to minimize the network connections subjected to the constraint of maximal agreement between the experimental and predicted gene dynamics. The developed algorithm is validated using both in silico and experimental data-sets. These studies show that the algorithm can accurately predict the topology and connection strength of the in silico networks, as quantified by high precision and recall, and small discrepancy between the actual and predicted kinetic parameters

Network motifs in a given network are small connected subnetworks that occur at significantly higher frequencies than would be expected for a random network. In their 2007 article “An optimalalgorithm for counting network motifs”, Itzhack, Mogilevski, and Louzoun present an algorithm for detecting network motifs. Based on an experimental comparison with a motif detection software called FANMOD, they claim that their algorithm is “more than a thousand times faster” than any previous motif detection algorithm. We show that this claim is not correct and based on a significant flaw in the experimental setup. Once the experimental data of Itzhack et al. is corrected for this flaw, the implementation of their algorithm actually turns out to be a little slower than FANMOD for random Erdős-Rényi graphs. For random scale-free networks, the implementation of Itzhack et al. is faster only by a factor of ˜1.5, not the orders of magnitude claimed by Itzhack et al.

Full Text Available Purpose: The purpose of this paper is to investigate the effectiveness of implementing unmanned aerial delivery vehicles in delivery networks. We investigate the notion of the reduced overall delivery time, energy, and costs for a truck-drone network by comparing the in-tandem system with a stand-alone delivery effort. The objectives are (1 to investigate the time, energy, and costs associated to a truck-drone delivery network compared to standalone truck or drone, (2 to propose an optimizationalgorithm that determines the optimal number of launch sites and locations given delivery requirements, and drones per truck, (3 to develop mathematical formulations for closed form estimations for the optimal number of launch locations, optimal total time, as well as the associated cost for the system. Design/methodology/approach: The design of the algorithm herein computes the minimal time of delivery utilizing K-means clustering to find launch locations, as well as a genetic algorithm to solve the truck route as a traveling salesmen problem (TSP. The optimal solution is determined by finding the minimum cost associated to the parabolic convex cost function. The optimal min-cost is determined by finding the most efficient launch locations using K-means algorithms to determine launch locations and a genetic algorithm to determine truck route between those launch locations. Findings: Results show improvements with in-tandem delivery efforts as opposed to standalone systems. Further, multiple drones per truck are more optimal and contribute to savings in both energy and time. For this, we sampled various initialization variables to derive closed form mathematical solutions for the problem. Originality/value: Ultimately, this provides the necessary analysis of an integrated truck-drone delivery system which could be implemented by a company in order to maximize deliveries while minimizing time and energy. Closed-form mathematical solutions can be used as

The purpose of this paper is to investigate the effectiveness of implementing unmanned aerial delivery vehicles in delivery networks. We investigate the notion of the reduced overall delivery time, energy, and costs for a truck-drone network by comparing the in-tandem system with a stand-alone delivery effort. The objectives are (1) to investigate the time, energy, and costs associated to a truck-drone delivery network compared to standalone truck or drone, (2) to propose an optimizationalgorithm that determines the optimal number of launch sites and locations given delivery requirements, and drones per truck, (3) to develop mathematical formulations for closed form estimations for the optimal number of launch locations, optimal total time, as well as the associated cost for the system. The design of the algorithm herein computes the minimal time of delivery utilizing K-means clustering to find launch locations, as well as a genetic algorithm to solve the truck route as a traveling salesmen problem (TSP). The optimal solution is determined by finding the minimum cost associated to the parabolic convex cost function. The optimal min-cost is determined by finding the most efficient launch locations using K-means algorithms to determine launch locations and a genetic algorithm to determine truck route between those launch locations. Results show improvements with in-tandem delivery efforts as opposed to standalone systems. Further, multiple drones per truck are more optimal and contribute to savings in both energy and time. For this, we sampled various initialization variables to derive closed form mathematical solutions for the problem. Ultimately, this provides the necessary analysis of an integrated truck-drone delivery system which could be implemented by a company in order to maximize deliveries while minimizing time and energy. Closed-form mathematical solutions can be used as close estimators for final costs and time. (Author)

Full Text Available To improve the lifetime of mobile sink-based wireless sensor networks and considering that data transmission delay and hops are limited in actual system, a lifetime optimizationalgorithm limited by data transmission delay and hops (LOA_DH for mobile sink-based wireless sensor networks is proposed. In LOA_DH, some constraints are analyzed, and an optimization model is proposed. Maximum capacity path routing algorithm is used to calculate the energy consumption of communication. Improved genetic algorithm which modifies individuals to meet all constraints is used to solve the optimization model. The optimal solution of sink node’s sojourn grid centers and sojourn times which maximizes network lifetime is obtained. Simulation results show that, in three node distribution scenes, LOA_DH can find the movement solution of sink node which covers all sensor nodes. Compared with MCP_RAND, MCP_GMRE, and EASR, the solution improves network lifetime and reduces average amount of node discarded data and average energy consumption of nodes.

In this article, a new optimization framework to reduce uncertainties in petroleum reservoir attributes using artificial intelligence techniques (neural network and genetic algorithm) is proposed. Instead of using the deterministic values of the reservoir properties, as in a conventional process, the parameters of the probability density function of each uncertain attribute are set as design variables in an optimization process using a genetic algorithm. The objective function (OF) is based on the misfit of a set of models, sampled from the probability density function, and a symmetry factor (which represents the distribution of curves around the history) is used as weight in the OF. Artificial neural networks are trained to represent the production curves of each well and the proxy models generated are used to evaluate the OF in the optimization process. The proposed method was applied to a reservoir with 16 uncertain attributes and promising results were obtained.

In this paper, we present a parallel test strategy for bandwidth division multiplexing under the test access mechanism bandwidth constraint. The Pareto solution set is combined with a cloud evolutionary algorithm to optimize the test time and power consumption of a three-dimensional network-on-chip (3D NoC). In the proposed method, all individuals in the population are sorted in non-dominated order and allocated to the corresponding level. Individuals with extreme and similar characteristics are then removed. To increase the diversity of the population and prevent the algorithm from becoming stuck around local optima, a competition strategy is designed for the individuals. Finally, we adopt an elite reservation strategy and update the individuals according to the cloud model. Experimental results show that the proposed algorithm converges to the optimal Pareto solution set rapidly and accurately. This not only obtains the shortest test time, but also optimizes the power consumption of the 3D NoC.

Full Text Available Purpose of the article: The ability of the company to predict customer churn and retain customers is considered to be worthy competitive advantage since it improves cost allocation in customer retention programs, retaining future revenue and profits. In addition, it has several positive indirect impacts such as increasing customer’s loyalty. Therefore, the focus of the article is on building highly reliable and robust classification model, which deals with such a task. Methodology/methods: The analysis is carried out on labelled ecommerce retail dataset describing 10 000 most valuable customers with the highest CLV (Customer Lifetime Value. To obtain the best performing ANN (Artificial Neural Network classification model, proposed hyperparameter search space is explored with genetic algorithm to find suitable parameter settings. ANN classification performance is measured with regard to prediction ability, which is understood as point estimate of AUC (Area Under Curve mean on 4fold cross-validation set. Explored part of hyperparameter search space is analyzed with conditional inference tree structure addressing underlying fundamental context of given optimization which results in identification of critical factors leading to well performing ANN classification model. Scientific aim: To present and execute experimental design for performance evaluation and hyperparameter optimization of classification models, which are used for customer churn prediction. Findings: It is concluded and statistically proven that in experimental context described, regularization parameter as well as training function have significant influence on classifiers AUC performance contrasting other properties of ANN. More specifically, well performing ANN classification models have regularization parameter set to 0, adaptation function set to trainlm or trainscg and more than 100 training epochs. Global optimum is identified for solution with regularization parameter set to

In this paper, a new model for MEPDN (multiyear expansion planning of distribution networks) is proposed. By solving this model, the optimal expansion scheme of primary (i.e. medium voltage) distribution network including the reinforcement pattern of primary feeders as well as location and size of DG (distributed generators) during an ascertained planning period is determined. Furthermore, the time-based feature of proposed model allows it to specify the investments/reinforcements time (i.e. year). Moreover, a minimum load shedding-based analytical approach for optimizing the network's reliability is introduced. The associated objective function of proposed model is minimizing the total investment and operation costs. To solve the formulated MEPDN model as a complex multi-dimensional optimization problem, a new evolutionary algorithm-based solution method called BCSSO (Binary Chaotic Shark Smell Optimization) is presented. The effectiveness of the proposed MEPDN model and solution approach is illustrated by applying them on two widely-used test cases including 12-bus and 33-bus distribution network and comparing the acquired results with the results of other solution methods. - Highlights: • A multiyear expansion planning model for distribution network is presented. • A new evolutionary algorithm-based solution approach is proposed. • A minimum load shedding-based analytical method for EENS minimization is suggested. • The efficacy of the proposed solution approach is broadly investigated.

The authors consider the algorithm of defining of basic parameters of a filter compensative device based on new element of electrical circuits—katkon. Katkon is a two-terminal network consisting of two parallel coiled and electrically unconnected coils having both inductive and capacitive properties. Filter compensative device based on katkon realizes reactive power compensation on the industrial frequency and filtering of high harmonics: in other words, it realizes functions of energy quality management in an electrical network. The design of such a filter compensative device and its equivalent circuit and algorithm of defining of its parameters that allows choosing its layout and design are described in the article. The results of physical experimental studies with katkon that confirm the adequacy of the proposed algorithm are demonstrated.

Highlights: • A detailed data processing will make more accurate results prediction. • Taking a full account of more load factors to improve the prediction precision. • Improved BP network obtains higher learning convergence. • Genetic algorithmoptimized by chaotic cat map enhances the global search ability. • The combined GA–BP model improved by modified additional momentum factor is superior to others. - Abstract: This paper proposes an appropriate combinational approach which is based on improved BP neural network for short-term gas load forecasting, and the network is optimized by the real-coded genetic algorithm. Firstly, several kinds of modifications are carried out on the standard neural network to accelerate the convergence speed of network, including improved additional momentum factor, improved self-adaptive learning rate and improved momentum and self-adaptive learning rate. Then, it is available to use the global search capability of optimized genetic algorithm to determine the initial weights and thresholds of BP neural network to avoid being trapped in local minima. The ability of GA is enhanced by cat chaotic mapping. In light of the characteristic of natural gas load for Shanghai, a series of data preprocessing methods are adopted and more comprehensive load factors are taken into account to improve the prediction accuracy. Such improvements facilitate forecasting efficiency and exert maximum performance of the model. As a result, the integration model improved by modified additional momentum factor gets more ideal solutions for short-term gas load forecasting, through analyses and comparisons of the above several different combinational algorithms

Transformers are essential devices of the power system. The accurate computation of the highest temperature (HST) of a transformer’s windings is very significant, as for the HST is a fundamental parameter in controlling the load operation mode and influencing the life time of the insulation. Based on the analysis of the heat transfer processes and the thermal characteristics inside transformers, there is taken into consideration the influence of factors like the sunshine, external wind speed etc. on the oil-immersed transformers. Experimental data and the neural network are used for modeling and protesting of the HST, and furthermore, investigations are conducted on the optimization of the structure and algorithms of neutral network are conducted. Comparison is made between the measured values and calculated values by using the recommended algorithm of IEC60076 and by using the neural networkalgorithm proposed by the authors; comparison that shows that the value computed with the neural networkalgorithm approximates better the measured value than the value computed with the algorithm proposed by IEC60076.

Three-dimensional heat transfer characteristics and pressure drop of water flow in a set of rectangular microchannels are numerically investigated using Fluent and compared with those of experimental results. Two metamodels based on the evolved group method of data handling (GMDH) type neural networks are then obtained for modelling of both pressure drop (ΔP) and Nusselt number (Nu) with respect to design variables such as geometrical parameters of microchannels, the amount of heat flux and the Reynolds number. Using such obtained polynomial neural networks, multi-objective genetic algorithms (GAs) (non-dominated sorting genetic algorithm, NSGA-II) with a new diversity preserving mechanism is then used for Pareto based optimization of microchannels considering two conflicting objectives such as (ΔP) and (Nu). It is shown that some interesting and important relationships as useful optimal design principles involved in the performance of microchannels can be discovered by Pareto based multi-objective optimization of the obtained polynomial metamodels representing their heat transfer and flow characteristics. Such important optimal principles would not have been obtained without the use of both GMDH type neural network modelling and the Pareto optimization approach

An optimization method for some of the CNN's (Cellular Neural Network) parameters, based on evolutionary strategies, is proposed. The new class of feedback template found is more effective in extracting features from the images that an autonomous vehicle acquires, than in the previous CNN's literature

Drying remains a complex process to model and thus to optimize. In this paper a new approach is proposed which allows to perform a compression in the drying model in order to integrate it using neural networks. The simulation times become very small and allow to test a great number of configurations. This decisive advantage allows to perform a multi-criteria optimization using hybrid genetical algorithms based on technical-economical criteria like drying cost, production or final product quality. (J.S.) 10 refs.

We present a new global optimizationalgorithm in which the influence of the leaders in social groups is used as an inspiration for the evolutionary technique which is designed into a group architecture. To demonstrate the efficiency of the method, a standard suite of single and multi-dimensional optimization functions along with the energies and the geometric structures of Lennard-Jones clusters are given as well as the application of the algorithm on quantum circuit design problems. We show that as an improvement over previous methods, the algorithm scales as N 2.5 for the Lennard-Jones clusters of N-particles. In addition, an efficient circuit design is shown for a two-qubit Grover search algorithm which is a quantum algorithm providing quadratic speedup over the classical counterpart.

Full Text Available Among artificial intelligence approaches, artificial neural networks (ANNs and genetic algorithm (GA are widely applied for modification of materials property in engineering science in large scale modeling. In this work artificial neural network (ANN and genetic algorithm (GA were applied to find the optimal conditions for achieving the maximum hardness of Al6061 reinforced by multiwall carbon nanotubes (MWCNTs through modeling of nanocomposite characteristics. After examination the different ANN architectures an optimal structure of the model, i.e. 6-18-1, is obtained with 1.52% mean absolute error and R2 = 0.987. The proposed structure was used as fitting function for genetic algorithm. The results of GA simulation predicted that the combination sintering temperature 346 °C, sintering time 0.33 h, compact pressure 284.82 MPa, milling time 19.66 h and vial speed 310.5 rpm give the optimum hardness, (i.e., 87.5 micro Vickers in the composite with 0.53 wt% CNT. Also, sensitivity analysis shows that the sintering time, milling time, compact pressure, vial speed and amount of MWCNT are the significant parameter and sintering time is the most important parameter. Comparison of the predicted values with the experimental data revealed that the GA–ANN model is a powerful method to find the optimal conditions for preparing of Al6061-MWCNT.

Full Text Available One of the main concerns of the mining industry is to determine ultimate pit limits. Final pit is a collection of blocks, which can be removed with maximum profit while following restrictions on the slope of the mine’s walls. The size, location and final shape of an open-pit are very important in designing the location of waste dumps, stockpiles, processing plants, access roads and other surface facilities as well as in developing a production program. There are numerous methods for designing ultimate pit limits. Some of these methods, such as floating cone algorithm, are heuristic and do not guarantee to generate optimum pit limits. Other methods, like Lerchs–Grossmann algorithm, are rigorous and always generate the true optimum pit limits. In this paper, a new rigorous algorithm is introduced. The main logic in this method is that only positive blocks, which can pay costs of their overlying non-positive blocks, are able to appear in the final pit. Those costs may be paid either by positive block itself or jointly with other positive blocks, which have the same overlying negative blocks. This logic is formulated using a network model as a Linear Programming (LP problem. This algorithm can be applied to two- and three-dimension block models. Since there are many commercial programs available for solving LP problems, pit limits in large block models can be determined easily by using this method.

Our aim was to predict tooth surface loss in individuals without the need to conduct clinical examinations. Artificial neural networks (ANNs) were used to construct a mathematical model. Input data consisted of age, smoker status, type of tooth brush, brushing, and consumption of pickled food, fizzy drinks, orange, apple, lemon, and dried seeds. Output data were the sum of tooth surface loss scores for selected teeth. The optimized constructed ANN consisted of 2-layer network with 15 neurons in the first layer and one neuron in the second layer. The data of 46 subjects were used to build the model, while the data of 15 subjects were used to test the model. Accepting an error of ±5 scores for all chosen teeth, the accuracy of the network becomes more than 80%. In conclusion, this study shows that modeling tooth surface loss using ANNs is possible and can be achieved with a high degree of accuracy.

Full Text Available Our aim was to predict tooth surface loss in individuals without the need to conduct clinical examinations. Artificial neural networks (ANNs were used to construct a mathematical model. Input data consisted of age, smoker status, type of tooth brush, brushing, and consumption of pickled food, fizzy drinks, orange, apple, lemon, and dried seeds. Output data were the sum of tooth surface loss scores for selected teeth. The optimized constructed ANN consisted of 2-layer network with 15 neurons in the first layer and one neuron in the second layer. The data of 46 subjects were used to build the model, while the data of 15 subjects were used to test the model. Accepting an error of ±5 scores for all chosen teeth, the accuracy of the network becomes more than 80%. In conclusion, this study shows that modeling tooth surface loss using ANNs is possible and can be achieved with a high degree of accuracy.

Full Text Available It is well known in the literature that the problem of learning the structure of Bayesian networks is very hard to tackle: Its computational complexity is super-exponential in the number of nodes in the worst case and polynomial in most real-world scenarios. Efficient implementations of score-based structure learning benefit from past and current research in optimization theory, which can be adapted to the task by using the network score as the objective function to maximize. This is not true for approaches based on conditional independence tests, called constraint-based learning algorithms. The only optimization in widespread use, backtracking, leverages the symmetries implied by the definitions of neighborhood and Markov blanket. In this paper we illustrate how backtracking is implemented in recent versions of the bnlearn R package, and how it degrades the stability of Bayesian network structure learning for little gain in terms of speed. As an alternative, we describe a software architecture and framework that can be used to parallelize constraint-based structure learning algorithms (also implemented in bnlearn and we demonstrate its performance using four reference networks and two real-world data sets from genetics and systems biology. We show that on modern multi-core or multiprocessor hardware parallel implementations are preferable over backtracking, which was developed when single-processor machines were the norm.

Full Text Available This paper proposes a new hybrid metaheuristique algorithm based on the hybridization of Biogeography-based optimization with the Differential Evolution for solving the optimal power flow problem with emission control. The biogeography-based optimization (BBO algorithm is strongly influenced by equilibrium theory of island biogeography, mainly through two steps: Migration and Mutation. Differential Evolution (DE is one of the best Evolutionary Algorithms for global optimization. The hybridization of these two methods is used to overcome traps of local optimal solutions and problems of time consumption. The objective of this paper is to minimize the total fuel cost of generation, total emission, total real power loss and also maintain an acceptable system performance in terms of limits on generator real power, bus voltages and power flow of transmission lines. In the present work, BBO/DE has been applied to solve the optimal power flow problems on IEEE 30-bus test system and the Algerian electrical network 114 bus. The results obtained from this method show better performances compared with DE, BBO and other well known metaheuristique and evolutionary optimization methods.

In this paper, an optimized fuzzy controller is presented for the control of the environmental parameters at the building zone level. The occupants' preferences are monitored via a smart card unit. Genetic algorithmoptimization techniques are applied to shift properly the membership functions of the fuzzy controller in order to satisfy the occupants' preferences while minimizing energy consumption. The implementation of the system integrates a smart card unit, sensors, actuators, interfaces, a programmable logic controller (PLC), local operating network (LON) modules and devices, and a central PC that monitors the performance of the system. The communication of the PLC with the smart card unit is performed using an RS 485 port, while the PLC-PC communication is performed via the LON network. The integrated system is installed and tested in the building of the Laboratory of Electronics of the Technical University of Crete. (author)

The optimalnetwork design problem has been well addressed in geodesy and photogrammetry but has not received the same attention for terrestrial laser scanner (TLS) networks. The goal of this research is to develop a complete design system that can automatically provide an optimal plan for high-accuracy, large-volume scanning networks. The aim in this paper is to use three heuristic optimization methods, simulated annealing (SA), genetic algorithm (GA) and particle swarm optimization (PSO), to solve the first-order design (FOD) problem for a small-volume indoor network and make a comparison of their performances. The room is simplified as discretized wall segments and possible viewpoints. Each possible viewpoint is evaluated with a score table representing the wall segments visible from each viewpoint based on scanning geometry constraints. The goal is to find a minimum number of viewpoints that can obtain complete coverage of all wall segments with a minimal sum of incidence angles. The different methods have been implemented and compared in terms of the quality of the solutions, runtime and repeatability. The experiment environment was simulated from a room located on University of Calgary campus where multiple scans are required due to occlusions from interior walls. The results obtained in this research show that PSO and GA provide similar solutions while SA doesn't guarantee an optimal solution within limited iterations. Overall, GA is considered as the best choice for this problem based on its capability of providing an optimal solution and fewer parameters to tune.

A framework is proposed for the design and analysis of network-oblivious algorithms, namely algorithms that can run unchanged, yet efficiently, on a variety of machines characterized by different degrees of parallelism and communication capabilities. The framework prescribes that a network-oblivi...

The improvement of high-through-put gene profiling based microarrays technology has provided monitoring the expression value of thousands of genes simultaneously. Detailed examination of changes in expression levels of genes can help physicians to have efficient diagnosing, classification of tumors and cancer's types as well as effective treatments. Finding genes that can classify the group of cancers correctly based on hybrid optimizationalgorithms is the main purpose of this paper. In this paper, a hybrid particle swarm optimization and genetic algorithm method are used for gene selection and also artificial neural network (ANN) is adopted as the classifier. In this work, we have improved the ability of the algorithm for the classification problem by finding small group of biomarkers and also best parameters of the classifier. The proposed approach is tested on three benchmark gene expression data sets: Blood (acute myeloid leukemia, acute lymphoblastic leukemia), colon and breast datasets. We used 10-fold cross-validation to achieve accuracy and also decision tree algorithm to find the relation between the biomarkers for biological point of view. To test the ability of the trained ANN models to categorize the cancers, we analyzed additional blinded samples that were not previously used for the training procedure. Experimental results show that the proposed method can reduce the dimension of the data set and confirm the most informative gene subset and improve classification accuracy with best parameters based on datasets.

In order to improve the measurement precision of 6-axis force/torque sensor for robot, BP decoupling algorithmoptimized by GA (GA-BP algorithm) is proposed in this paper. The weights and thresholds of a BP neural network with 6-10-6 topology are optimized by GA to develop decouple a six-axis force/torque sensor. By comparison with other traditional decoupling algorithm, calculating the pseudo-inverse matrix of calibration and classical BP algorithm, the decoupling results validate the good decoupling performance of GA-BP algorithm and the coupling errors are reduced.

Full Text Available In wireless sensor networks, data aggregation routing could reduce the number of data transmissions so as to achieve energy efficient transmission. However, data aggregation introduces data retransmission that is caused by co-channel interference from neighboring sensor nodes. This kind of co-channel interference could result in extra energy consumption and significant latency from retransmission. This will jeopardize the benefits of data aggregation. One possible solution to circumvent data retransmission caused by co-channel interference is to assign different channels to every sensor node that is within each other’s interference range on the data aggregation tree. By associating each radio with a different channel, a sensor node could receive data from all the children nodes on the data aggregation tree simultaneously. This could reduce the latency from the data source nodes back to the sink so as to meet the user’s delay QoS. Since the number of radios on each sensor node and the number of non-overlapping channels are all limited resources in wireless sensor networks, a challenging question here is to minimize the total transmission cost under limited number of non-overlapping channels in multi-radio wireless sensor networks. This channel constrained data aggregation routing problem in multi-radio wireless sensor networks is an NP-hard problem. I first model this problem as a mixed integer and linear programming problem where the objective is to minimize the total transmission subject to the data aggregation routing, channel and radio resources constraints. The solution approach is based on the Lagrangean relaxation technique to relax some constraints into the objective function and then to derive a set of independent subproblems. By optimally solving these subproblems, it can not only calculate the lower bound of the original primal problem but also provide useful information to get the primal feasible solutions. By incorporating these

Full Text Available Due to their special environment, Underwater Wireless Sensor Networks (UWSNs are usually deployed over a large sea area and the nodes are usually floating. This results in a lower beacon node distribution density, a longer time for localization, and more energy consumption. Currently most of the localization algorithms in this field do not pay enough consideration on the mobility of the nodes. In this paper, by analyzing the mobility patterns of water near the seashore, a localization method for UWSNs based on a Mobility Prediction and a Particle Swarm Optimizationalgorithm (MP-PSO is proposed. In this method, the range-based PSO algorithm is used to locate the beacon nodes, and their velocities can be calculated. The velocity of an unknown node is calculated by using the spatial correlation of underwater object’s mobility, and then their locations can be predicted. The range-based PSO algorithm may cause considerable energy consumption and its computation complexity is a little bit high, nevertheless the number of beacon nodes is relatively smaller, so the calculation for the large number of unknown nodes is succinct, and this method can obviously decrease the energy consumption and time cost of localizing these mobile nodes. The simulation results indicate that this method has higher localization accuracy and better localization coverage rate compared with some other widely used localization methods in this field.

Due to their special environment, Underwater Wireless Sensor Networks (UWSNs) are usually deployed over a large sea area and the nodes are usually floating. This results in a lower beacon node distribution density, a longer time for localization, and more energy consumption. Currently most of the localization algorithms in this field do not pay enough consideration on the mobility of the nodes. In this paper, by analyzing the mobility patterns of water near the seashore, a localization method for UWSNs based on a Mobility Prediction and a Particle Swarm Optimizationalgorithm (MP-PSO) is proposed. In this method, the range-based PSO algorithm is used to locate the beacon nodes, and their velocities can be calculated. The velocity of an unknown node is calculated by using the spatial correlation of underwater object's mobility, and then their locations can be predicted. The range-based PSO algorithm may cause considerable energy consumption and its computation complexity is a little bit high, nevertheless the number of beacon nodes is relatively smaller, so the calculation for the large number of unknown nodes is succinct, and this method can obviously decrease the energy consumption and time cost of localizing these mobile nodes. The simulation results indicate that this method has higher localization accuracy and better localization coverage rate compared with some other widely used localization methods in this field.

Highlights: • Decoupling of a heating, ventilation, and air conditioning system is presented. • RBF models were identified by Epsilon constraint method for temperature and humidity. • Control settings derived from optimization of the decoupled model. • Epsilon constraint-RBF based on PID controller was implemented to keep thermal comfort and minimize energy. • Enhancements of controller parameters of the HVAC system are desired. - Abstract: The energy efficiency of a heating, ventilating and air conditioning (HVAC) system optimized using a radial basis function neural network (RBFNN) combined with the epsilon constraint (EC) method is reported. The new method adopts the advanced algorithm of RBFNN for the HVAC system to estimate the residual errors, increase the control signal and reduce the error results. The objective of this study is to develop and simulate the EC-RBFNN for a self tuning PID controller for a decoupled bilinear HVAC system to control the temperature and relative humidity (RH) produced by the system. A case study indicates that the EC-RBFNN algorithm has a much better accuracy than optimization PID itself and PID-RBFNN, respectively.

This work presents a prototype of a system for nuclear reactor core design optimization based on genetic algorithms and artificial neural networks. A neural network is modeled and trained in order to predict the flux and the neutron multiplication factor values based in the enrichment, network pitch and cladding thickness, with average error less than 2%. The values predicted by the neural network are used by a genetic algorithm in this heuristic search, guided by an objective function that rewards the high flux values and penalizes multiplication factors far from the required value. Associating the quick prediction - that may substitute the reactor physics calculation code - with the global optimization capacity of the genetic algorithm, it was obtained a quick and effective system for nuclear reactor core design optimization. (author). 11 refs., 8 figs., 3 tabs.

This study proposes three methodologies to define artificial neural network models through genetic algorithms (GAs) to predict the next-day hourly average surface ozone (O(3)) concentrations. GAs were applied to define the activation function in hidden layer and the number of hidden neurons. Two of the methodologies define threshold models, which assume that the behaviour of the dependent variable (O(3) concentrations) changes when it enters in a different regime (two and four regimes were considered in this study). The change from one regime to another depends on a specific value (threshold value) of an explanatory variable (threshold variable), which is also defined by GAs. The predictor variables were the hourly average concentrations of carbon monoxide (CO), nitrogen oxide, nitrogen dioxide (NO(2)), and O(3) (recorded in the previous day at an urban site with traffic influence) and also meteorological data (hourly averages of temperature, solar radiation, relative humidity and wind speed). The study was performed for the period from May to August 2004. Several models were achieved and only the best model of each methodology was analysed. In threshold models, the variables selected by GAs to define the O(3) regimes were temperature, CO and NO(2) concentrations, due to their importance in O(3) chemistry in an urban atmosphere. In the prediction of O(3) concentrations, the threshold model that considers two regimes was the one that fitted the data most efficiently.

Subgraph matching algorithms are designed to find all instances of predefined subgraphs in a large graph or network and play an important role in the discovery and analysis of so-called network motifs, subgraph patterns which occur more often than expected by chance. We present the index-based subgraph matching algorithm (ISMA), a novel tree-based algorithm. ISMA realizes a speedup compared to existing algorithms by carefully selecting the order in which the nodes of a query subgraph are investigated. In order to achieve this, we developed a number of data structures and maximally exploited symmetry characteristics of the subgraph. We compared ISMA to a naive recursive tree-based algorithm and to a number of well-known subgraph matching algorithms. Our algorithm outperforms the other algorithms, especially on large networks and with large query subgraphs. An implementation of ISMA in Java is freely available at http://sourceforge.net/projects/isma/.

Full Text Available Subgraph matching algorithms are designed to find all instances of predefined subgraphs in a large graph or network and play an important role in the discovery and analysis of so-called network motifs, subgraph patterns which occur more often than expected by chance. We present the index-based subgraph matching algorithm (ISMA, a novel tree-based algorithm. ISMA realizes a speedup compared to existing algorithms by carefully selecting the order in which the nodes of a query subgraph are investigated. In order to achieve this, we developed a number of data structures and maximally exploited symmetry characteristics of the subgraph. We compared ISMA to a naive recursive tree-based algorithm and to a number of well-known subgraph matching algorithms. Our algorithm outperforms the other algorithms, especially on large networks and with large query subgraphs. An implementation of ISMA in Java is freely available at http://sourceforge.net/projects/isma/.

This paper presents the development of multi-objective Genetic Algorithms to optimize chlorination design and management in drinking water networks (DWN). Three objectives have been considered: the improvement of the chlorination uniformity (healthy objective), the minimization of chlorine booster stations number, and the injected chlorine mass (economic objectives). The problem has been dissociated in medium and short terms ones. The proposed methodology was tested on hypothetical and real DWN. Results proved the ability of the developed optimization tool to identify relationships between the healthy and economic objectives as Pareto fronts. The proposed approach was efficient in computing solutions ensuring better chlorination uniformity while requiring the weakest injected chlorine mass when compared to other approaches. For the real DWN studied, chlorination optimization has been crowned by great improvement of free-chlorine-dosing uniformity and by a meaningful chlorine mass reduction, in comparison with the conventional chlorination.

Full Text Available For predicting the key technology indicators (concentrate grade and tailings recovery rate of flotation process, a feed-forward neural network (FNN based soft-sensor model optimized by the hybrid algorithm combining particle swarm optimization (PSO algorithm and gravitational search algorithm (GSA is proposed. Although GSA has better optimization capability, it has slow convergence velocity and is easy to fall into local optimum. So in this paper, the velocity vector and position vector of GSA are adjusted by PSO algorithm in order to improve its convergence speed and prediction accuracy. Finally, the proposed hybrid algorithm is adopted to optimize the parameters of FNN soft-sensor model. Simulation results show that the model has better generalization and prediction accuracy for the concentrate grade and tailings recovery rate to meet the online soft-sensor requirements of the real-time control in the flotation process.

Candidate network designs tested for acceptability and cost. OptimalNetwork Topology Design computer program developed as part of study on topology design and analysis of performance of Space Station Information System (SSIS) network. Uses efficient algorithm to generate candidate network designs consisting of subsets of set of all network components, in increasing order of total costs and checks each design to see whether it forms acceptable network. Technique gives true cost-optimalnetwork and particularly useful when network has many constraints and not too many components. Program written in PASCAL.

Full Text Available Our ultimate goal is to design transportation net- works whose dynamic performance metrics (e.g. pas- senger throughput, passenger delay, and insensitivity to weather disturbances are optimized. Here the fo- cus is on optimizing static features of the network that are known to directly aﬀect the network dynamics. First, we present simulation results which support a connection between maximizing the ﬁrst non-trivial eigenvalue of a network's Laplacian and superior air- port network performance. Then, we explore the ef- fectiveness of a tabu search heuristic for optimizing this metric by comparing experimental results to the- oretical upper bounds. We also consider generating upper bounds on a network's algebraic connectivity via the solution of semideﬁnite programming (SDP relaxations. A modiﬁcation of an existing subgraph extraction algorithm is implemented to explore the underlying regional structures in the U.S. airport net- work, with the hope that the resulting localized struc- tures can be optimized independently and reconnected via a "backbone" network to achieve superior network performance.

Short-segment instrumentation for spine fractures is threatened by relatively high failure rates. Failure of the spinal pedicle screws including breakage and loosening may jeopardize the fixation integrity and lead to treatment failure. Two important design objectives, bending strength and pullout strength, may conflict with each other and warrant a multiobjective optimization study. In the present study using the three-dimensional finite element (FE) analytical results based on an L25 orthogonal array, bending and pullout objective functions were developed by an artificial neural network (ANN) algorithm, and the trade-off solutions known as Pareto optima were explored by a genetic algorithm (GA). The results showed that the knee solutions of the Pareto fronts with both high bending and pullout strength ranged from 92% to 94% of their maxima, respectively. In mechanical validation, the results of mathematical analyses were closely related to those of experimental tests with a correlation coefficient of -0.91 for bending and 0.93 for pullout (P design had significantly higher fatigue life (P < 0.01) and comparable pullout strength as compared with commercial screws. Multiobjective optimization study of spinal pedicle screws using the hybrid of ANN and GA could achieve an ideal with high bending and pullout performances simultaneously.

Describes how evolutionary algorithms (EAs) can be used to identify, model, and minimize day-to-day problems that arise for researchers in optimization and mobile networking. Mobile ad hoc networks (MANETs), vehicular networks (VANETs), sensor networks (SNs), and hybrid networks—each of these require a designer’s keen sense and knowledge of evolutionary algorithms in order to help with the common issues that plague professionals involved in optimization and mobile networking. This book introduces readers to both mobile ad hoc networks and evolutionary algorithms, presenting basic concepts as well as detailed descriptions of each. It demonstrates how metaheuristics and evolutionary algorithms (EAs) can be used to help provide low-cost operations in the optimization process—allowing designers to put some “intelligence” or sophistication into the design. It also offers efficient and accurate information on dissemination algorithms topology management, and mobility models to address challenges in the ...

Full Text Available The water footprint of energy systems must be considered, as future water scarcity has been identified as a major concern. This work presents a general life cycle network modeling and optimization framework for energy-based products and processes using a functional unit of liters of water consumed in the processing pathway. We analyze and optimize the water-energy nexus over the objectives of water footprint minimization, maximization of economic output per liter of water consumed (economic efficiency of water, and maximization of energy output per liter of water consumed (energy efficiency of water. A mixed integer, multiobjective nonlinear fractional programming (MINLFP model is formulated. A mixed integer linear programing (MILP-based branch and refine algorithm that incorporates both the parametric algorithm and nonlinear programming (NLP subproblems is developed to boost solving efficiency. A case study in bioenergy is presented, and the water footprint is considered from biomass cultivation to biofuel production, providing a novel perspective into the consumption of water throughout the value chain. The case study, optimized successively over the three aforementioned objectives, utilizes a variety of candidate biomass feedstocks to meet primary fuel products demand (ethanol, diesel, and gasoline. A minimum water footprint of 55.1 ML/year was found, economic efficiencies of water range from −$1.31/L to $0.76/L, and energy efficiencies of water ranged from 15.32 MJ/L to 27.98 MJ/L. These results show optimization provides avenues for process improvement, as reported values for the energy efficiency of bioethanol range from 0.62 MJ/L to 3.18 MJ/L. Furthermore, the proposed solution approach was shown to be an order of magnitude more efficient than directly solving the original MINLFP problem with general purpose solvers.

Transmission network expansion planning (TNEP) is a basic part of power system planning that determines where, when and how many new transmission lines should be added to the network. Up till now, various methods have been presented to solve the static transmission network expansion planning (STNEP) problem. But in all of these methods, transmission expansion planning considering network adequacy restriction has not been investigated. Thus, in this paper, STNEP problem is...

Many physicists are not aware of the fact that they can solve their problems by applying optimizationalgorithms. Since the number of such algorithms is steadily increasing, many new algorithms have not been presented comprehensively until now. This presentation of recently developed algorithms applied in physics, including demonstrations of how they work and related results, aims to encourage their application, and as such the algorithms selected cover concepts and methods from statistical physics to optimization problems emerging in theoretical computer science.

A selective method based on molecularly imprinted polymer (MIP) solid-phase extraction (SPE) using UV-Vis spectrophotometry as a detection technique was developed for the determination of fluoxetine (FLU) in pharmaceutical and human serum samples. The MIPs were synthesized using pyrrole as a functional monomer in the presence of FLU as a template molecule. The factors that affecting the preparation and extraction ability of MIP such as amount of sorbent, initiator concentration, the amount of monomer to template ratio, uptake shaking rate, uptake time, washing buffer pH, take shaking rate, Taking time and polymerization time were considered for optimization. First a Plackett-Burman design (PBD) consists of 12 randomized runs were applied to determine the influence of each factor. The other optimization processes were performed using central composite design (CCD), artificial neural network (ANN) and genetic algorithm (GA). At optimal condition the calibration curve showed linearity over a concentration range of 10- 7-10- 8 M with a correlation coefficient (R2) of 0.9970. The limit of detection (LOD) for FLU was obtained 6.56 × 10- 9 M. The repeatability of the method was obtained 1.61%. The synthesized MIP sorbent showed a good selectivity and sensitivity toward FLU. The MIP/SPE method was used for the determination of FLU in pharmaceutical, serum and plasma samples, successfully.

This thesis develops a new method to detect delaminations in composite laminates using a combination of finite element method, artificial neural networks, and genetic algorithms. Next, this newly developed method is applied to successfully solve delamination detection problems. Delaminations in a composite laminate with various sizes and locations are considered in the present studies. The improved layerwise shear deformation theory is implemented into the finite element method and used to calculate responses of laminates with single and multiple delaminations. Mappings between the natural frequencies and delamination characteristics are first determined from the developed models. These data are then used to train artificial neural networks of multiplayer perceptron using back-propagation. These trained artificial neural networks are in turn used as an approximate tool to calculate the responses of the delaminated laminates and to feed the data to the delamination detection process. Two different approaches for handling the neural network models are applied in the work and are presented for comparison. The delamination detection problem is formulated as an optimization problem with mixed type design variables. A genetic algorithm, which is a guided probabilistic search technique based on the simulation of Darwin's principle of evolution and natural selection, is developed to solve this optimization problem. Single through-the-width delamination, single internal delamination, and multiple through-the-width delaminations are separately considered for detection study. At last, the application is extended to the most challenging problem, which is the detection of general delamination. Various factors affecting the detection process such as the finite element convergence factor and the laminate geometry factor are also examined. Case studies are made and the findings are summarized in detail in each chapter of the dissertation. It is found that the newly developed

htmlabstractThe study of phylogenetic networks is of great interest to computational evolutionary biology and numerous different types of such structures are known. This article addresses the following question concerning rooted versions of phylogenetic networks. What is the maximum value of pset

Full Text Available It is everyone’s dream to have network connectivity anywhere at all times. This dream can only be realized provided there are feasible solutions that are put in place for the next generation of wireless works. Wireless Mesh Networks (WMNs...

Full Text Available The intermittency of renewable energy will increase the uncertainty of the power system, so it is necessary to predict the short-term wind power, after which the electrical power system can operate reliably and safely. Unlike the traditional point forecasting, the purpose of this study is to quantify the potential uncertainties of wind power and to construct prediction intervals (PIs and prediction models using wavelet neural network (WNN. Lower upper bound estimation (LUBE of the PIs is achieved by minimizing a multi-objective function covering both interval width and coverage probabilities. Considering the influence of the points out of the PIs to shorten the width of PIs without compromising coverage probability, a new, improved, multi-objective artificial bee colony (MOABC algorithm combining multi-objective evolutionary knowledge, called EKMOABC, is proposed for the optimization of the forecasting model. In this paper, some comparative simulations are carried out and the results show that the proposed model and algorithm can achieve higher quality PIs for wind power forecasting. Taking into account the intermittency of renewable energy, such a type of wind power forecast can actually provide a more reliable reference for dispatching of the power system.

Hypoglycemia or low blood glucose is dangerous and can result in unconsciousness, seizures and even death. It is a common and serious side effect of insulin therapy in patients with diabetes. HypoMon is a non-invasive monitor that measures some physiological parameters continuously to provide detection of hypoglycemic episodes in Type 1 diabetes mellitus patients (T1DM). Based on heart rate and corrected QT interval of the ECG signal, we have continued to develop Bayesian neural network detection algorithms to recognize the presence of hypoglycemic episodes. From a clinical study of 16 children with T1DM, natural occurrence of nocturnal hypoglycemic episodes are associated with increased heart rates (1.033+/-0.242 vs. 1.082+/-0.298, P0.06) and increased corrected QT intervals (1.031+/-0.086 vs. 1.060+/-0.084, P0.001). The overall data were organized into a training set (8 patients) and a test set (another 8 patients) randomly selected. Using the optimal Bayesian neural network with 10 hidden nodes which was derived from the training set with the highest log evidence, the sensitivity (true positive) value for detection of hypoglycemia in the test set is 89.2%.

Full Text Available Abstract In this study a newly developed thin-walled structure with the combination of circular and square sections is investigated in term of crashworthiness. The results of the experimental tests are utilized to validate the Abaqus/ExplicitTM finite element simulations and analysis of the crush phenomenon. Three polynomial meta-models based on the evolved group method of data handling (GMDH neural networks are employed to simply represent the specific energy absorption (SEA, the initial peak crushing load (P1 and the secondary peak crushing load (P2 with respect to the geometrical variables. The training and testing data are extracted from the finite element analysis. The modified genetic algorithm NSGA-II, is used in multi-objective optimisation of the specific energy absorption, primary and secondary peak crushing load according to the geometrical variables. Finally, in each optimisation process, the optimal section energy absorptions are compared with the results of the finite element analysis. The nearest to ideal point and TOPSIS optimisation methods are applied to choose the optimal points.

We present an optimizationalgorithm to obtain low-uncertainty dynamic pressure measurements from a force-transducer-based device. In this paper, the advantages and disadvantages of the methods that are commonly used to measure the propellant powder gas pressure, the applicable scope of dynamic pressure calibration devices, and the shortcomings of the traditional comparison calibration method based on the drop-weight device are firstly analysed in detail. Then, a dynamic calibration method for measuring pressure using a force sensor based on a drop-weight device is introduced. This method can effectively save time when many pressure sensors are calibrated simultaneously and extend the life of expensive reference sensors. However, the force sensor is installed between the drop-weight and the hammerhead by transition pieces through the connection mode of bolt fastening, which causes adverse effects such as additional pretightening and inertia forces. To solve these effects, the influence mechanisms of the pretightening force, the inertia force and other influence factors on the force measurement are theoretically analysed. Then a measurement correction method for the force measurement is proposed based on an artificial neural networkoptimized by a genetic algorithm. The training and testing data sets are obtained from calibration tests, and the selection criteria for the key parameters of the correction model is discussed. The evaluation results for the test data show that the correction model can effectively improve the force measurement accuracy of the force sensor. Compared with the traditional high-accuracy comparison calibration method, the percentage difference of the impact-force-based measurement is less than 0.6% and the relative uncertainty of the corrected force value is 1.95%, which can meet the requirements of engineering applications.

energy. An example of this is when an aerial vehicle like a quadrotor is used as a communications relay station. Since quadrotors are like helicopters...then the quadrotors may be driven to land somewhere, which may be suboptimal in terms of network metrics, but is much better in terms of energy

Long-term groundwater monitoring networks can provide essential information for the planning and management of water resources. Budget constraints in water resource management agencies often mean a reduction in the number of observation wells included in a monitoring network. A network design tool, distributed as an R package, was developed to determine which wells to exclude from a monitoring network because they add little or no beneficial information. A kriging-based genetic algorithm method was used to optimize the monitoring network. The algorithm was used to find the set of wells whose removal leads to the smallest increase in the weighted sum of the (1) mean standard error at all nodes in the kriging grid where the water table is estimated, (2) root-mean-squared-error between the measured and estimated water-level elevation at the removed sites, (3) mean standard deviation of measurements across time at the removed sites, and (4) mean measurement error of wells in the reduced network. The solution to the optimization problem (the best wells to retain in the monitoring network) depends on the total number of wells removed; this number is a management decision. The network design tool was applied to optimize two observation well networks monitoring the water table of the eastern Snake River Plain aquifer, Idaho; these networks include the 2008 Federal-State Cooperative water-level monitoring network (Co-op network) with 166 observation wells, and the 2008 U.S. Geological Survey-Idaho National Laboratory water-level monitoring network (USGS-INL network) with 171 wells. Each water-level monitoring network was optimized five times: by removing (1) 10, (2) 20, (3) 40, (4) 60, and (5) 80 observation wells from the original network. An examination of the trade-offs associated with changes in the number of wells to remove indicates that 20 wells can be removed from the Co-op network with a relatively small degradation of the estimated water table map, and 40 wells

as a sequential game . First, the defender prepares for an attack by either hardening portions of a system or, in some cases, designing new additions...sequential game in which the decisions are made in a predetermined sequence: the defender, then the attacker and finally the defender (or operator...held constant while the remaining variables are optimized. We define the hat symbol (^) to denote the holding of a decision variable to a fixed value

In this research, MgAl-CO 3 2- nanolayered double hydroxide (NLDH) was synthesized through a facile coprecipitation method, followed by a hydrothermal treatment. The prepared NLDHs were used as a hydrophilic nanofiller for improving the performance of the PVDF-based ultrafiltration membranes. The main objective of this research was to obtain the optimized formula of NLDH/PVDF nanocomposite membrane presenting the best performance using computational techniques as a cost-effective method. For this aim, an artificial neural network (ANN) model was developed for modeling and expressing the relationship between the performance of the nanocomposite membrane (pure water flux, protein flux and flux recovery ratio) and the affecting parameters including the NLDH, PVP 29000 and polymer concentrations. The effects of the mentioned parameters and the interaction between the parameters were investigated using the contour plot predicted with the developed model. Scanning electron microscopy (SEM), atomic force microscopy (AFM), and water contact angle techniques were applied to characterize the nanocomposite membranes and to interpret the predictions of the ANN model. The developed ANN model was introduced to genetic algorithm (GA) as a bioinspired optimizer to determine the optimum values of input parameters leading to high pure water flux, protein flux, and flux recovery ratio. The optimum values for NLDH, PVP 29000 and the PVDF concentration were determined to be 0.54, 1, and 18 wt %, respectively. The performance of the nanocomposite membrane prepared using the optimum values proposed by GA was investigated experimentally, in which the results were in good agreement with the values predicted by ANN model with error lower than 6%. This good agreement confirmed that the nanocomposite membranes prformance could be successfully modeled and optimized by ANN-GA system.

Full Text Available We investigate a new method to place patients into risk groups in censored survival data. Properties such as median survival time, and end survival rate, are implicitly improved by optimizing the area under the survival curve. Artificial neural networks (ANN are trained to either maximize or minimize this area using a genetic algorithm, and combined into an ensemble to predict one of low, intermediate, or high risk groups. Estimated patient risk can influence treatment choices, and is important for study stratification. A common approach is to sort the patients according to a prognostic index and then group them along the quartile limits. The Cox proportional hazards model (Cox is one example of this approach. Another method of doing risk grouping is recursive partitioning (Rpart, which constructs a decision tree where each branch point maximizes the statistical separation between the groups. ANN, Cox, and Rpart are compared on five publicly available data sets with varying properties. Cross-validation, as well as separate test sets, are used to validate the models. Results on the test sets show comparable performance, except for the smallest data set where Rpart's predicted risk groups turn out to be inverted, an example of crossing survival curves. Cross-validation shows that all three models exhibit crossing of some survival curves on this small data set but that the ANN model manages the best separation of groups in terms of median survival time before such crossings. The conclusion is that optimizing the area under the survival curve is a viable approach to identify risk groups. Training ANNs to optimize this area combines two key strengths from both prognostic indices and Rpart. First, a desired minimum group size can be specified, as for a prognostic index. Second, the ability to utilize non-linear effects among the covariates, which Rpart is also able to do.

Boolean networks are widely used to model gene regulatory networks and to design therapeutic intervention strategies to affect the long-term behavior of systems. In this paper, we investigate the less-studied one-bit perturbation, which falls under the category of structural intervention. Previous works focused on finding the optimal one-bit perturbation to maximally alter the steady-state distribution (SSD) of undesirable states through matrix perturbation theory. However, the application of the SSD is limited to Boolean networks with about ten genes. In 2007, Xiao et al. proposed to search the optimal one-bit perturbation by altering the sizes of the basin of attractions (BOAs). However, their algorithm requires close observation of the state-transition diagram. In this paper, we propose an algorithm that efficiently determines the BOA size after a perturbation. Our idea is that, if we construct the basin of states for all states, then the size of the BOA of perturbed networks can be obtained just by updating the paths of the states whose transitions have been affected. Results from both synthetic and real biological networks show that the proposed algorithm performs better than the exhaustive SSD-based algorithm and can be applied to networks with about 25 genes.

Full Text Available This paper presents application of a new effective metaheuristic optimization method namely, the Jaya algorithm to deal with different optimum power flow (OPF problems. Unlike other population-based optimization methods, no algorithm-particular controlling parameters are required for this algorithm. In this work, three goal functions are considered for the OPF solution: generation cost minimization, real power loss reduction, and voltage stability improvement. In addition, the effect of distributed generation (DG is incorporated into the OPF problem using a modified formulation. For best allocation of DG unit(s, a sensitivity-based procedure is introduced. Simulations are carried out on the modified IEEE 30-bus and IEEE 118-bus networks to determine the effectiveness of the Jaya algorithm. The single objective optimization cases are performed both with and without DG. For all considered cases, results demonstrate that Jaya algorithm can produce an optimum solution with rapid convergence. Statistical analysis is also carried out to check the reliability of the Jaya algorithm. The optimal solution obtained by the Jaya algorithm is compared with different stochastic algorithms, and demonstrably outperforms them in terms of solution optimality and solution feasibility, proving its effectiveness and potential. Notably, optimal placement of DGs results in even better solutions.

Perceptively written text examines optimization problems that can be formulated in terms of networks and algebraic structures called matroids. Chapters cover shortest paths, network flows, bipartite matching, nonbipartite matching, matroids and the greedy algorithm, matroid intersections, and the matroid parity problems. A suitable text or reference for courses in combinatorial computing and concrete computational complexity in departments of computer science and mathematics.

Traffic monitoring and managing in urban intelligent transportation systems (ITS) can be carried out based on vehicular sensor networks. In a vehicular sensor network, vehicles equipped with sensors such as GPS, can act as mobile sensors for sensing the urban traffic and sending the reports to a traffic monitoring center (TMC) for traffic estimation. The energy consumption by the sensor nodes is a main problem in the wireless sensor networks (WSNs); moreover, it is the most important feature in designing these networks. Clustering the sensor nodes is considered as an effective solution to reduce the energy consumption of WSNs. Each cluster should have a Cluster Head (CH), and a number of nodes located within its supervision area. The cluster heads are responsible for gathering and aggregating the information of clusters. Then, it transmits the information to the data collection center. Hence, the use of clustering decreases the volume of transmitting information, and, consequently, reduces the energy consumption of network. In this paper, Fuzzy C-Means (FCM) and Fuzzy Subtractive algorithms are employed to cluster sensors and investigate their performance on the energy consumption of sensors. It can be seen that the FCM algorithm and Fuzzy Subtractive have been reduced energy consumption of vehicle sensors up to 90.68% and 92.18%, respectively. Comparing the performance of the algorithms implies the 1.5 percent improvement in Fuzzy Subtractive algorithm in comparison.

Full Text Available Traffic monitoring and managing in urban intelligent transportation systems (ITS can be carried out based on vehicular sensor networks. In a vehicular sensor network, vehicles equipped with sensors such as GPS, can act as mobile sensors for sensing the urban traffic and sending the reports to a traffic monitoring center (TMC for traffic estimation. The energy consumption by the sensor nodes is a main problem in the wireless sensor networks (WSNs; moreover, it is the most important feature in designing these networks. Clustering the sensor nodes is considered as an effective solution to reduce the energy consumption of WSNs. Each cluster should have a Cluster Head (CH, and a number of nodes located within its supervision area. The cluster heads are responsible for gathering and aggregating the information of clusters. Then, it transmits the information to the data collection center. Hence, the use of clustering decreases the volume of transmitting information, and, consequently, reduces the energy consumption of network. In this paper, Fuzzy C-Means (FCM and Fuzzy Subtractive algorithms are employed to cluster sensors and investigate their performance on the energy consumption of sensors. It can be seen that the FCM algorithm and Fuzzy Subtractive have been reduced energy consumption of vehicle sensors up to 90.68% and 92.18%, respectively. Comparing the performance of the algorithms implies the 1.5 percent improvement in Fuzzy Subtractive algorithm in comparison.

This program was developed as part of a research study on the topology design and performance analysis for the Space Station Information System (SSIS) network. It uses an efficient algorithm to generate candidate network designs (consisting of subsets of the set of all network components) in increasing order of their total costs, and checks each design to see if it forms an acceptable network. This technique gives the true cost-optimalnetwork, and is particularly useful when the network has many constraints and not too many components. It is intended that this new design technique consider all important performance measures explicitly and take into account the constraints due to various technical feasibilities. In the current program, technical constraints are taken care of by the user properly forming the starting set of candidate components (e.g. nonfeasible links are not included). As subsets are generated, they are tested to see if they form an acceptable network by checking that all requirements are satisfied. Thus the first acceptable subset encountered gives the cost-optimal topology satisfying all given constraints. The user must sort the set of "feasible" link elements in increasing order of their costs. The program prompts the user for the following information for each link: 1) cost, 2) connectivity (number of stations connected by the link), and 3) the stations connected by that link. Unless instructed to stop, the program generates all possible acceptable networks in increasing order of their total costs. The program is written only to generate topologies that are simply connected. Tests on reliability, delay, and other performance measures are discussed in the documentation, but have not been incorporated into the program. This program is written in PASCAL for interactive execution and has been implemented on an IBM PC series computer operating under PC DOS. The disk contains source code only. This program was developed in 1985.

Full Text Available The problem of finding the optimal distribution of pressure drop over a network is solved via an unconstrained gradient type algorithm. The developed algorithm is computationally attractive. Problems with several hundred variables and constraints were solved.

This book provides a comprehensive account of the glowworm swarm optimization (GSO) algorithm, including details of the underlying ideas, theoretical foundations, algorithm development, various applications, and MATLAB programs for the basic GSO algorithm. It also discusses several research problems at different levels of sophistication that can be attempted by interested researchers. The generality of the GSO algorithm is evident in its application to diverse problems ranging from optimization to robotics. Examples include computation of multiple optima, annual crop planning, cooperative exploration, distributed search, multiple source localization, contaminant boundary mapping, wireless sensor networks, clustering, knapsack, numerical integration, solving fixed point equations, solving systems of nonlinear equations, and engineering design optimization. The book is a valuable resource for researchers as well as graduate and undergraduate students in the area of swarm intelligence and computational intellige...

Full Text Available In this article, the cutting parameters optimization method for aluminum alloy AlMn1Cu in high-speed milling was studied in order to properly select the high-speed cutting parameters. First, a back propagation neural network model for predicting surface roughness of AlMn1Cu was proposed. The prediction model can improve the prediction accuracy and well work out the higher-order nonlinear relationship between surface roughness and cutting parameters. Second, considering the constraints of technical requirements on surface roughness, a mathematical model for optimizing cutting parameters based on the Bayesian neural network prediction model of surface roughness was established so as to obtain the maximum machining efficiency. The genetic algorithm adopting the homogeneous design to initialize population as well as steady-state reproduction without duplicates was also presented. The application indicates that the algorithm can effectively avoid precocity, strengthen global optimization, and increase the calculation efficiency. Finally, a case was presented on the application of the proposed cutting parameters optimizationalgorithm to optimize the cutting parameters.

Are you looking for new lectures for your course on algorithms, combinatorial optimization, or algorithmic game theory? Maybe you need a convenient source of relevant, current topics for a graduate student or advanced undergraduate student seminar? Or perhaps you just want an enjoyable look at some beautiful mathematical and algorithmic results, ideas, proofs, concepts, and techniques in discrete mathematics and theoretical computer science? Gems of Combinatorial Optimization and Graph Algorithms is a handpicked collection of up-to-date articles, carefully prepared by a select group of international experts, who have contributed some of their most mathematically or algorithmically elegant ideas. Topics include longest tours and Steiner trees in geometric spaces, cartograms, resource buying games, congestion games, selfish routing, revenue equivalence and shortest paths, scheduling, linear structures in graphs, contraction hierarchies, budgeted matching problems, and motifs in networks. This ...

Optimizing and controlling of complex engineering systems is a phenomenon that has attracted an incremental interest of numerous scientists. Until now, a variety of intelligent optimizing and controlling techniques such as neural networks, fuzzy logic, game theory, support vector machines...... performance of the proposed method. In order to find the maximum exploration potentials, these techniques are equipped with an external archive. These archives aid the methods to record all of the non-dominated solutions. Eventually, the proposed method and generalized regression neural network (GRNN......) are simultaneously used to optimize the major parameters of an irreversible thermal engine. In order to direct the PBMSB to explore deliberate spaces within the solution domain, a reference point obtained from finite time thermodynamic (FTT) approach, is utilized in the optimization. The outcome results show...

Understanding the emergence of cooperation has long been a challenge across disciplines. Even if network reciprocity reflected the importance of population structure in promoting cooperation, it remains an open question how population structures can be optimized, thereby enhancing cooperation. In this paper, we attempt to apply the evolutionary algorithm (EA) to solve this highly complex problem. However, as it is hard to evaluate the fitness (cooperation level) of population structures, simply employing the canonical evolutionary algorithm (EA) may fail in optimization. Thus, we propose a new EA variant named mlEA-C PD -SFN to promote the cooperation level of scale-free networks (SFNs) in the Prisoner's Dilemma Game (PDG). Meanwhile, to verify the preceding conclusions may not be applied to this problem, we also provide the optimization results of the comparative experiment (EA cluster ), which optimizes the clustering coefficient of structures. Even if preceding research concluded that highly clustered scale-free networks enhance cooperation, we find EA cluster does not perform desirably, while mlEA-C PD -SFN performs efficiently in different optimization environments. We hope that mlEA-C PD -SFN may help promote the structure of species in nature and that more general properties that enhance cooperation can be learned from the output structures.

Network science is a rapidly emerging field of study that encompasses mathematics, computer science, physics, and engineering. A key issue in the study of complex networks is to understand the collective behavior of the various elements of these networks.Although the results from graph theory have proven to be powerful in investigating the structures of complex networks, few books focus on the algorithmic aspects of complex network analysis. Filling this need, Complex Networks: An Algorithmic Perspective supplies the basic theoretical algorithmic and graph theoretic knowledge needed by every r

Hypoglycemia or low blood glucose is a common and serious side effect of insulin therapy in patients with diabetes. HypoMon is a non-invasive monitor that measures some physiological parameters continuously to provide detection of hypoglycemic episodes in Type 1 diabetes mellitus patients (T1DM). Based on heart rate, corrected QT interval of the ECG signal and skin impedance, a Bayesian neural network detection algorithm has been developed to recognize the presence of hypoglycemic episodes. From a clinical study of 25 children with T1DM, associated with hypoglycemic episodes, their heart rates increased (1.152+/-0.157 vs. 1.035+/-0.108, PBayesian neural network with 11 hidden nodes, and an algorithm developed from the training set, a sensitivity of 0.8346 and specificity of 0.6388 were achieved for the test set.

Full Text Available Inspired by the process of migration and reproduction of flora, this paper proposes a novel artificial flora (AF algorithm. This algorithm can be used to solve some complex, non-linear, discrete optimization problems. Although a plant cannot move, it can spread seeds within a certain range to let offspring to find the most suitable environment. The stochastic process is easy to copy, and the spreading space is vast; therefore, it is suitable for applying in intelligent optimizationalgorithm. First, the algorithm randomly generates the original plant, including its position and the propagation distance. Then, the position and the propagation distance of the original plant as parameters are substituted in the propagation function to generate offspring plants. Finally, the optimal offspring is selected as a new original plant through the selection function. The previous original plant becomes the former plant. The iteration continues until we find out optimal solution. In this paper, six classical evaluation functions are used as the benchmark functions. The simulation results show that proposed algorithm has high accuracy and stability compared with the classical particle swarm optimization and artificial bee colony algorithm.

The main reason for the efficiency decrease at part load conditions for four-stroke spark-ignition (SI) engines is the flow restriction at the cross-sectional area of the intake system. Traditionally, valve-timing has been designed to optimize operation at high engine-speed and wide open throttle conditions. Several investigations have demonstrated that improvements at part load conditions in engine performance can be accomplished if the valve-timing is variable. Controlling valve-timing can be used to improve the torque and power curve as well as to reduce fuel consumption and emissions. In this paper, a group method of data handling (GMDH) type neural network and evolutionary algorithms (EAs) are firstly used for modelling the effects of intake valve-timing (V t ) and engine speed (N) of a spark-ignition engine on both developed engine torque (T) and fuel consumption (Fc) using some experimentally obtained training and test data. Using such obtained polynomial neural network models, a multi-objective EA (non-dominated sorting genetic algorithm, NSGA-II) with a new diversity preserving mechanism are secondly used for Pareto based optimization of the variable valve-timing engine considering two conflicting objectives such as torque (T) and fuel consumption (Fc). The comparison results demonstrate the superiority of the GMDH type models over feedforward neural network models in terms of the statistical measures in the training data, testing data and the number of hidden neurons. Further, it is shown that some interesting and important relationships, as useful optimal design principles, involved in the performance of the variable valve-timing four-stroke spark-ignition engine can be discovered by the Pareto based multi-objective optimization of the polynomial models. Such important optimal principles would not have been obtained without the use of both the GMDH type neural network modelling and the multi-objective Pareto optimization approach

Previous experiments have shown that fungi use an efficient natural algorithm for searching the space available for their growth in micro-confined networks, e.g., mazes. This natural "master" algorithm, which comprises two "slave" sub-algorithms, i.e., collision-induced branching and directional memory, has been shown to be more efficient than alternatives, with one, or the other, or both sub-algorithms turned off. In contrast, the present contribution compares the performance of the fungal natural algorithm against several standard artificial homologues. It was found that the space-searching fungal algorithm consistently outperforms uninformed algorithms, such as Depth-First-Search (DFS). Furthermore, while the natural algorithm is inferior to informed ones, such as A*, this under-performance does not importantly increase with the increase of the size of the maze. These findings suggest that a systematic effort of harvesting the natural space searching algorithms used by microorganisms is warranted and possibly overdue. These natural algorithms, if efficient, can be reverse-engineered for graph and tree search strategies.

Full Text Available This contribution deals with an application of three different algorithms which optimize a group delay of analog filters. One of them is a part of the professional circuit simulator Micro Cap 7 and the others two original algorithms are developed in the MATLAB environment. An all-pass network is used to optimize the group delay of an arbitrary analog filter. Introduced algorithms look for an optimal order and optimal coefficients of an all-pass network transfer function. Theoretical foundations are introduced and all algorithms are tested using the optimization of testing analog filter. The optimization is always run three times because the second, third and fourth-order all-pass network is used. An equalization of the original group delay is a main objective of these optimizations. All outputs of all algorithms are critically evaluated and also described.

Earthquake early warning (EEW) systems should provide reliable warnings as quickly as possible with a minimum number of false and missed alarms. Using the example of the megacity Istanbul and based on a set of simulated scenario earthquakes, we present a novel approach for evaluating and optimizing seismic networks for EEW, in particular in regions with a scarce number of instrumentally recorded earthquakes. We show that, while the current station locations of the existing Istanbul EEW system...

Full Text Available We propose a timing-driven discrete cell-sizing algorithm that can address total cell size and/or leakage power constraints. We model cell sizing as a “discretized” mincost network flow problem, wherein available sizes of each cell are modeled as nodes. Flow passing through a node indicates the choice of the corresponding cell size, and the total flow cost reflects the timing objective function value corresponding to these choices. Compared to other discrete optimization methods for cell sizing, our method can obtain near-optimal solutions in a time-efficient manner. We tested our algorithm on ISCAS’85 benchmarks, and compared our results to those produced by an optimal dynamic programming- (DP- based method. The results show that compared to the optimal method, the improvements to an initial sizing solution obtained by our method is only 1% (3% worse when using a 180 nm (90 nm library, while being 40–60 times faster. We also obtained results for ISPD’12 cell-sizing benchmarks, under leakage power constraint, and compared them to those of a state-of-the-art approximate DP method (optimal DP runs out of memory for the smallest of these circuits. Our results show that we are only 0.9% worse than the approximate DP method, while being more than twice as fast.

Proportional Integral Derivative (PID) controllers are widely used in industrial processes for their simplicity and robustness. The main application problems are the tuning of PID parameters to obtain good settling time, rise time and overshoot. The challenge is to improve the timing parameters to achieve optimal control performances. Remarkable findings are obtained through the use of Artificial Intelligence techniques as Fuzzy Logic, Genetic Algorithms and Neural Networks. The combination o...

Full Text Available Accurate and stable prediction of icing thickness on transmission lines is of great significance for ensuring the safe operation of the power grid. In order to improve the accuracy and stability of icing prediction, an innovative prediction model based on the generalized regression neural network (GRNN and the fruit fly optimizationalgorithm (FOA is proposed. Firstly, a feature selection method based on the data inconsistency rate (IR is adopted to select the optimal feature, which aims to reduce redundant input vectors. Then, the fruit FOA is utilized for optimization of smoothing factor for the GRNN. Lastly, the icing forecasting method FOA-IR-GRNN is established. Two cases in different locations and different months are selected to validate the proposed model. The results indicate that the new hybrid FOA-IR-GRNN model presents better accuracy, robustness, and generality in icing forecasting.

This study addresses the optimization of node redeployment coverage in underwater wireless sensor networks. Given that nodes could easily become invalid under a poor environment and the large scale of underwater wireless sensor networks, an underwater sensor network redeployment algorithm was developed based on wolf search. This study is to apply the wolf search algorithm combined with crowded degree control in the deployment of underwater wireless sensor networks. The proposed algorithm uses nodes to ensure coverage of the events, and it avoids the prematurity of the nodes. The algorithm has good coverage effects. In addition, considering that obstacles exist in the underwater environment, nodes are prevented from being invalid by imitating the mechanism of avoiding predators. Thus, the energy consumption of the network is reduced. Comparative analysis shows that the algorithm is simple and effective in wireless sensor network deployment. Compared with the optimized artificial fish swarm algorithm, the proposed algorithm exhibits advantages in network coverage, energy conservation, and obstacle avoidance.

Full Text Available Artificial neural networks are complex networks emulating the way human rational neurons process data. They have been widely used generally in prediction clustering classification and association. The training algorithms that used to determine the network weights are almost the most important factor that influence the neural networks performance. Recently many meta-heuristic and Evolutionary algorithms are employed to optimize neural networks weights to achieve better neural performance. This paper aims to use recently proposed algorithms for optimizing neural networks weights comparing these algorithms performance with other classical meta-heuristic algorithms used for the same purpose. However to evaluate the performance of such algorithms for training neural networks we examine such algorithms to classify four opposite binary XOR clusters and classification of continuous real data sets such as Iris and Ecoli.

Full Text Available In order to accurately evaluate the health condition of a shearer, a hybrid prediction method was proposed based on the integration of a fuzzy neural network (FNN and improved particle swarm optimization (IPSO. The parameters of FNN were optimized by the use of PSO, which was coupled with a premature judgment and mutation mechanism to increase the convergence speed and enhance the generalization ability. The key technologies are elaborated and the flowchart of the proposed approach was designed. Furthermore, an experiment example was carried out and the comparison results indicated that the proposed approach was feasible and outperforms others. Finally, a field application example in coal mining face was demonstrated to specify the effect of the proposed system.

Network design is by nature costly and optimization models play significant role in reducing the unnecessary cost components of a distribution network. This study proposes a genetic algorithm to solve a distribution network design model. The structure of the chromosome in the proposed algorithm is defined in a novel way that in addition to producing feasible solutions, it also reduces the computational complexity of the algorithm. Computational results are presented to show the algorithm performance.

Genetic algorithm (GA) inspired by Darwin's theory of evolution and employed to solve optimization problems - unconstrained or constrained - uses an evolutionary process. A GA has several parameters such the population size, search space, crossover and mutation probabilities, and fitness criterion. These parameters are not universally known/determined a priori for all problems. Depending on the problem at hand, these parameters need to be decided such that the resulting GA performs the best. We present here a preprocessor that achieves just that, i.e., it determines, for a specified problem, the foregoing parameters so that the consequent GA is a best for the problem. We stress also the need for such a preprocessor both for quality (error) and for cost (complexity) to produce the solution. The preprocessor includes, as its first step, making use of all the information such as that of nature/character of the function/system, search space, physical/laboratory experimentation (if already done/available), and the physical environment. It also includes the information that can be generated through any means - deterministic/nondeterministic/graphics. Instead of attempting a solution of the problem straightway through a GA without having/using the information/knowledge of the character of the system, we would do consciously a much better job of producing a solution by using the information generated/created in the very first step of the preprocessor. We, therefore, unstintingly advocate the use of a preprocessor to solve a real-world optimization problem including NP-complete ones before using the statistically most appropriate GA. We also include such a GA for unconstrained function optimization problems.

Full Text Available Introduction The diagnosis of agricultural machinery faults must be performed at an opportune time, in order to fulfill the agricultural operations in a timely manner and to optimize the accuracy and the integrity of a system, proper monitoring and fault diagnosis of the rotating parts is required. With development of fault diagnosis methods of rotating equipment, especially bearing failure, the security, performance and availability of machines has been increasing. In general, fault detection is conducted through a specific procedure which starts with data acquisition and continues with features extraction, and subsequently failure of the machine would be detected. Several practical methods have been introduced for fault detection in rotating parts of machineries. The review of the literature shows that both Artificial Neural Networks (ANN and Support Vector Machines (SVM have been used for this purpose. However, the results show that SVM is more effective than Artificial Neural Networks in fault detection of such machineries. In some smart detection systems, incorporating an optimized method such as Genetic Algorithm in the Neural Network model, could improve the fault detection procedure. Consequently, the fault detection performance of neural networks may also be improved by combining with the Genetic Algorithm and hence will be comparable with the performance of the Support Vector Machine. In this study, the so called Genetic Algorithm (GA method was used to optimize the structure of the Artificial Neural Networks (ANN for fault detection of the clutch retainer mechanism of Massey Ferguson 285 tractor. Materials and Methods The test rig consists of some electro mechanical parts including the clutch retainer mechanism of Massey Ferguson 285 tractor, a supporting shaft, a single-phase electric motor, a loading mechanism to model the load of the tractor clutch and the corresponding power train gears. The data acquisition section consists of a

Today, tracking the growing interest in closed-loop supply chain shown by both practitioners and academia is easily possible. There are many factors, which transform closed-loop supply chain issues into a unique and vital subject in supply chain management, such as environmental legislation...... is proposed and a complete validation process is undertaken using CPLEX and MATLAB software. In small instances, the global optimum points of CPLEX for the proposed hybrid algorithm are compared to genetic algorithm, and particle swarm optimization. Then, in small, mid, and large-size instances, performances...

Full Text Available Reducing power losses in energy distribution is forced by mandatory law. Optimization of network reconfiguration can lead to reducing power losses by even a dozen or so percent. Two methods of network reconfiguration optimization are presented in the paper: the heuristic method and a method based on genetic algorithms. The presented solutions offer the opportunity to optimize medium voltage and low voltage distribution networks at the same time. Calculation results for real distribution network are presented. The presented results indicate a high efficiency of network reconfiguration optimization.

particle swarm optimization-sliding mode control (PSO-SMC) theory-based hybrid algorithm is proposed for the training of T2-FNNs. In particular, continuous version of PSO is adopted for the identification of the antecedent part of T2-FNNs while SMCbased update rules are utilized for online learning......, an autonomous quality inspection over rice farms is proposed by employing quadcopters. Real-time control of these vehicles, however, is still challenging as they exhibit highly nonlinear behavior especially for agile maneuvers. What is more, these vehicles have to operate under uncertain working conditions...... such as wind and gust disturbances as well as positioning errors caused by inertial measurement units and global positioning system. To handle these difficulties, as a model-free and learning control algorithm, type-2 fuzzy neural networks (T2-FNNs) are designed for the control of quadcopter. The novel...

This comprehensive textbook on combinatorial optimization places special emphasis on theoretical results and algorithms with provably good performance, in contrast to heuristics. It is based on numerous courses on combinatorial optimization and specialized topics, mostly at graduate level. This book reviews the fundamentals, covers the classical topics (paths, flows, matching, matroids, NP-completeness, approximation algorithms) in detail, and proceeds to advanced and recent topics, some of which have not appeared in a textbook before. Throughout, it contains complete but concise proofs, and also provides numerous exercises and references. This sixth edition has again been updated, revised, and significantly extended. Among other additions, there are new sections on shallow-light trees, submodular function maximization, smoothed analysis of the knapsack problem, the (ln 4+ɛ)-approximation for Steiner trees, and the VPN theorem. Thus, this book continues to represent the state of the art of combinatorial opti...

Methylobacillus sp. zju323 was adopted to improve the biosynthesis of pyrroloquinoline quinone (PQQ) by systematic optimization of the fermentation medium. The Plackett-Burman design was implemented to screen for the key medium components for the PQQ production. CoCl 2 · 6H 2 O, ρ-amino benzoic acid, and MgSO 4 · 7H 2 O were found capable of enhancing the PQQ production most significantly. A five-level three-factor central composite design was used to investigate the direct and interactive effects of these variables. Both response surface methodology (RSM) and artificial neural network-genetic algorithm (ANN-GA) were used to predict the PQQ production and to optimize the medium composition. The results showed that the medium optimized by ANN-GA was better than that by RSM in maximizing PQQ production and the experimental PQQ concentration in the ANN-GA-optimized medium was improved by 44.3% compared with that in the unoptimized medium. Further study showed that this ANN-GA-optimized medium was also effective in improving PQQ production by fed-batch mode, reaching the highest PQQ accumulation of 232.0 mg/L, which was about 47.6% increase relative to that in the original medium. The present work provided an optimized medium and developed a fed-batch strategy which might be potentially applicable in industrial PQQ production.

Multi objective optimization of externalities of traffic is performed solving a network design problem in which Dynamic Traffic Management measures are used. The resulting Pareto optimal set is determined by employing the SPEA2+ evolutionary algorithm.

Recently security issues like investment and financing in China's power industry have become increasingly prominent, bringing serious challenges to the financial security of the domestic power industry. Thus, it deserves to develop financial safety evaluation towards the Chinese power industry and is of practical significance. In this paper, the GA (genetic algorithm) is used to optimize the connection weights and thresholds of the traditional BPNN (back propagation neural network) so the new model of BPNN based on GA is established, hereinafter referred to as GA-BPNN (back propagation neural network based on genetic algorithm). Then, an empirical example of the electric power industry in China during the period 2003–2010 was selected to verify the proposed algorithm. By comparison with three other algorithms, the results indicate the model can be applied to evaluate the financial security of China's power industry effectively. Then index values of the financial security of China's power industry in 2011 were obtained according to the tested prediction model and the comprehensive safety scores and grades are calculated by the weighted algorithm. Finally, we analyzed the reasons and throw out suggestions based on the results. The work of this paper will provide a reference for the financial security evaluation of the energy industry in the future. - Highlights: • GA-BPNN model is applied to assess the financial security of China's power industry. • 12 indexes of 3 major categories are selected to build the evaluation index system. • The GA-BPNN is superior to the models of GM (1,1), BPNN and LSSVM on the whole. • Predicted financial safety status of China's power industry in 2011 is basic safe. • Reasons and suggestions are proposed based on the forecast results.

This paper proposes an artificial immune system (AIS) algorithm approach for reconfiguring distribution network with the presence distributed generators (DG). The distribution network with high-performance is a network that has a low power loss, better voltage profile, and loading balance among feeders. The task for improving the performance of the distribution network is optimization of network configuration. The optimization has become a necessary study with the presence of DG in entire networks. In this work, optimization of network configuration is based on an AIS algorithm. The methodology has been tested in a model of 33 bus IEEE radial distribution networks with and without DG integration. The results have been showed that the optimal configuration of the distribution network is able to reduce power loss and to improve the voltage profile of the distribution network significantly.

Strictly evolutionary approaches to improving the air transport system a highly complex network of interacting systems no longer suffice in the face of demand that is projected to double or triple in the near future. Thus evolutionary approaches should be augmented with active design methods. The ability to actively design, optimize and control a system presupposes the existence of predictive modeling and reasonably well-defined functional dependences between the controllable variables of the system and objective and constraint functions for optimization. Following recent advances in the studies of the effects of network topology structure on dynamics, we investigate the performance of dynamic processes on transport networks as a function of the first nontrivial eigenvalue of the network's Laplacian, which, in turn, is a function of the network s connectivity and modularity. The last two characteristics can be controlled and tuned via optimization. We consider design optimization problem formulations. We have developed a flexible simulation of network topology coupled with flows on the network for use as a platform for computational experiments.

Full Text Available Total soluble solids (TSS and fruit fresh weight are two indicators to show the quality of tomato fruits. To gain high values of TSS and fruit fresh weight, it is important to consider the concentration of nutrient solution, which is commonly represented by Electrical Conductivity (EC value. Generally, the increasing of EC value not only increases the number of TSS, but also decreases fruit fresh weight. Therefore, it is important to optimize the EC value for both indicators of quality of tomato fruits. The objective of this research is to optimize the EC value of nutrient solution on each generative stage using Artificial Neural Network (ANN and Genetic Algorithms (GA. ANN was used to identify the relationship between different EC value treatments with TSS value and fruit fresh weight. GA was applied to determine the optimal EC value in generative growth, which is divided into three stages. Results showed that the optimal EC values in the flowering stage, the fruiting stage and the harvesting stage were 1.4 mS/cm, 10.2 mS/cm and 9.7 mS/cm, respectively. Using these values, a tomato fruit could be estimated with TSS value of 7.9% and fruit fresh weight of 51.34 g.

Full Text Available Abstract Background Drug therapy has become increasingly efficient, with more drugs available for treatment of an ever-growing number of conditions. Yet, drug use is reported to be sub optimal in several aspects, such as dosage, patient's adherence and outcome of therapy. The aim of the current study was to investigate the possibility to optimize drug therapy using computer programs, available on the Internet. Methods One hundred and ten officially endorsed text documents, published between 1996 and 2004, containing guidelines for drug therapy in 246 disorders, were analyzed with regard to information about patient-, disease- and drug-related factors and relationships between these factors. This information was used to construct algorithms for identifying optimum treatment in each of the studied disorders. These algorithms were categorized in order to define as few models as possible that still could accommodate the identified factors and the relationships between them. The resulting program prototypes were implemented in HTML (user interface and JavaScript (program logic. Results Three types of algorithms were sufficient for the intended purpose. The simplest type is a list of factors, each of which implies that the particular patient should or should not receive treatment. This is adequate in situations where only one treatment exists. The second type, a more elaborate model, is required when treatment can by provided using drugs from different pharmacological classes and the selection of drug class is dependent on patient characteristics. An easily implemented set of if-then statements was able to manage the identified information in such instances. The third type was needed in the few situations where the selection and dosage of drugs were depending on the degree to which one or more patient-specific factors were present. In these cases the implementation of an established decision model based on fuzzy sets was required. Computer programs

In this paper, we propose an optimizationalgorithm based on the intelligent behavior of stem cell swarms in reproduction and self-organization. Optimizationalgorithms, such as the Genetic Algorithm (GA), Particle Swarm Optimization (PSO) algorithm, Ant Colony Optimization (ACO) algorithm and Artificial Bee Colony (ABC) algorithm, can give solutions to linear and non-linear problems near to the optimum for many applications; however, in some case, they can suffer from becoming trapped in local optima. The Stem Cells Algorithm (SCA) is an optimizationalgorithm inspired by the natural behavior of stem cells in evolving themselves into new and improved cells. The SCA avoids the local optima problem successfully. In this paper, we have made small changes in the implementation of this algorithm to obtain improved performance over previous versions. Using a series of benchmark functions, we assess the performance of the proposed algorithm and compare it with that of the other aforementioned optimizationalgorithms. The obtained results prove the superiority of the Modified Stem Cells Algorithm (MSCA).

Chaos optimizationalgorithms as a novel method of global optimization have attracted much attention, which were all based on Logistic map. However, we have noticed that the probability density function of the chaotic sequences derived from Logistic map is a Chebyshev-type one, which may affect the global searching capacity and computational efficiency of chaos optimizationalgorithms considerably. Considering the statistical property of the chaotic sequences of Logistic map and Kent map, the improved hybrid chaos-BFGS optimizationalgorithm and the Kent map based hybrid chaos-BFGS algorithm are proposed. Five typical nonlinear functions with multimodal characteristic are tested to compare the performance of five hybrid optimizationalgorithms, which are the conventional Logistic map based chaos-BFGS algorithm, improved Logistic map based chaos-BFGS algorithm, Kent map based chaos-BFGS algorithm, Monte Carlo-BFGS algorithm, mesh-BFGS algorithm. The computational performance of the five algorithms is compared, and the numerical results make us question the high efficiency of the chaos optimizationalgorithms claimed in some references. It is concluded that the efficiency of the hybrid optimizationalgorithms is influenced by the statistical property of chaotic/stochastic sequences generated from chaotic/stochastic algorithms, and the location of the global optimum of nonlinear functions. In addition, it is inappropriate to advocate the high efficiency of the global optimizationalgorithms only depending on several numerical examples of low-dimensional functions

This paper applies back propagation neural network (BPNN) optimized by genetic algorithm (GA) for the prediction of pressure generated by a drop-weight device and the quasi-static calibration of piezoelectric high-pressure sensors for the measurement of propellant powder gas pressure. The method can effectively overcome the slow convergence and local minimum problems of BPNN. Based on test data of quasi-static comparison calibration method, a mathematical model between each parameter of drop-weight device and peak pressure and pulse width was established, through which the practical quasi-static calibration without continuously using expensive reference sensors could be realized. Compared with multiple linear regression method, the GA-BPNN model has higher prediction accuracy and stability. The percentages of prediction error of peak pressure and pulse width are less than 0.7% and 0.3%, respectively.

Before the appearance of broadband links and wireless systems, networks have been used to connect people in new ways. Now, the modern world is connected through large-scale, computational networked systems such as the Internet. Because of the ever-advancing technology of networking, efficient algorithms have become increasingly necessary to solve some of the problems developing in this area. "Mathematical Aspects of Network Routing Optimization" focuses on computational issues arising from the process of optimizingnetwork routes, such as quality of the resulting links and their reli

Since the introduction of genetic algorithms in the 1970s, an enormous number of articles together with several significant monographs and books have been published on this methodology. As a result, genetic algorithms have made a major contribution to optimization, adaptation, and learning in a wide variety of unexpected fields. Over the years, many excellent books in genetic algorithmoptimization have been published; however, they focus mainly on single-objective discrete or other hard optimization problems under certainty. There appears to be no book that is designed to present genetic algorithms for solving not only single-objective but also fuzzy and multiobjective optimization problems in a unified way. Genetic Algorithms And Fuzzy Multiobjective Optimization introduces the latest advances in the field of genetic algorithmoptimization for 0-1 programming, integer programming, nonconvex programming, and job-shop scheduling problems under multiobjectiveness and fuzziness. In addition, the book treats a w...

This contributed volume offers a collection of papers presented at the 2016 Network Games, Control, and Optimization conference (NETGCOOP), held at the University of Avignon in France, November 23-25, 2016. These papers highlight the increasing importance of network control and optimization in many networking application domains, such as mobile and fixed access networks, computer networks, social networks, transportation networks, and, more recently, electricity grids and biological networks. Covering a wide variety of both theoretical and applied topics in the areas listed above, the authors explore several conceptual and algorithmic tools that are needed for efficient and robust control operation, performance optimization, and better understanding the relationships between entities that may be acting cooperatively or selfishly in uncertain and possibly adversarial environments. As such, this volume will be of interest to applied mathematicians, computer scientists, engineers, and researchers in other relate...

Real-time information is important for travelers' routing decisions in uncertain networks by enabling online adaptation to revealed traffic conditions. Usually there are spatial and/or temporal limitations in traveler information. In this research, a...

In optimizing the particle swarm optimization (PSO) that inevitable existence problem of prematurity and the local convergence, this paper base on this aspects is put forward a kind of modified particle swarm optimizationalgorithm, take the gradient descent method (BP algorithm) as a particle swarm operator embedded in particle swarm algorithm, and at the same time use to attenuation wall (Damping) approach to make fly off the search area of the particles of size remain unchanged and avoid t...

Full Text Available This paper investigates a solution of the optimization problem concerning the construction of diameter-optimal regular networks (graphs. Regular networks are of practical interest as the graph-theoretical models of reliable communication networks of parallel supercomputer systems, as a basis of the structure in a model of small world in optical and neural networks. It presents a new class of parametrically described regular networks - hypercirculant networks (graphs. An approach that uses evolutionary algorithms for the automatic generation of parametric descriptions of optimal hypercirculant networks is developed. Synthesis of optimal hypercirculant networks is based on the optimal circulant networks with smaller degree of nodes. To construct optimal hypercirculant networks is used a template of circulant network from the known optimal families of circulant networks with desired number of nodes and with smaller degree of nodes. Thus, a generating set of the circulant network is used as a generating subset of the hypercirculant network, and the missing generators are synthesized by means of the evolutionary algorithm, which is carrying out minimization of diameter (average diameter of networks. A comparative analysis of the structural characteristics of hypercirculant, toroidal, and circulant networks is conducted. The advantage hypercirculant networks under such structural characteristics, as diameter, average diameter, and the width of bisection, with comparable costs of the number of nodes and the number of connections is demonstrated. It should be noted the advantage of hypercirculant networks of dimension three over four higher-dimensional tori. Thus, the optimization of hypercirculant networks of dimension three is more efficient than the introduction of an additional dimension for the corresponding toroidal structures. The paper also notes the best structural parameters of hypercirculant networks in comparison with iBT-networks previously

Full Text Available In the fields which require finding the most appropriate value, optimization became a vital approach to employ effective solutions. With the use of optimization techniques, many different fields in the modern life have found solutions to their real-world based problems. In this context, classical optimization techniques have had an important popularity. But after a while, more advanced optimization problems required the use of more effective techniques. At this point, Computer Science took an important role on providing software related techniques to improve the associated literature. Today, intelligent optimization techniques based on Artificial Intelligence are widely used for optimization problems. The objective of this paper is to provide a comparative study on the employment of classical optimization solutions and Artificial Intelligence solutions for enabling readers to have idea about the potential of intelligent optimization techniques. At this point, two recently developed intelligent optimizationalgorithms, Vortex OptimizationAlgorithm (VOA and Cognitive Development OptimizationAlgorithm (CoDOA, have been used to solve some multidisciplinary optimization problems provided in the source book Thomas' Calculus 11th Edition and the obtained results have compared with classical optimization solutions.

Full Text Available Modern real world science and engineering problems can be classified as multi-objective optimisation problems which demand for expedient and efficient stochastic algorithms to respond to the optimization needs. This paper presents an object-oriented software application that implements a firework optimizationalgorithm for function optimization problems. The algorithm, a kind of parallel diffuse optimizationalgorithm is based on the explosive phenomenon of fireworks. The algorithm presented promising results when compared to other population or iterative based meta-heuristic algorithm after it was experimented on five standard benchmark problems. The software application was implemented in Java with interactive interface which allow for easy modification and extended experimentation. Additionally, this paper validates the effect of runtime on the algorithm performance.

A multivariate method based on solvent terminated dispersive liquid-liquid microextraction was developed for the determination of Cu 2+ ions in aqueous samples. In the proposed approach, di-2-ethylhexylphosphoric acid, xylene and acetone were used as chelating agent, dispersive and extraction solvents, respectively. The effects of various factors on the extraction efficiency such as extraction and dispersive solvent volumes, salt addition and pH were studied using central composite design (CCD) and artificial neural networks coupled bees algorithm (ANN-BA). Upon comparison of these techniques, ANN-BA model was considered to be better optimization method due to its higher percentage relative recovery (about 5%) as compared to the CCD approach. The linear range and the limits of detection (S/N = 3) and quantitation (S/N = 10) were 0.22-140, 0.08 and 0.22 µg L -1 , respectively. Under the optimal conditions, the recoveries for real samples spiked with 0.1 and 0.3 mg L -1 were in the range of 85-98%.

Full Text Available It is important to determine the seepage field parameters of a high core rockfill dam using the seepage data obtained during operation. For the Nuozhadu high core rockfill dam, a back analysis model is proposed using the radial basis function neural networkoptimized using a particle swarm optimizationalgorithm (PSO-RBFNN and the technology of finite element analysis for solving the saturated-unsaturated seepage field. The recorded osmotic pressure curves of osmometers, which are distributed in the maximum cross section, are applied to this back analysis. The permeability coefficients of the dam materials are retrieved using the measured seepage pressure values while the steady state seepage condition exists; that is, the water lever remains unchanged. Meanwhile, the parameters are tested using the unstable saturated-unsaturated seepage field while the water level rises. The results show that the permeability coefficients are reasonable and can be used to study the real behavior of a seepage field of a high core rockfill dam during its operation period.

The efficiency of a hybrid systems method which combined artificial neural networks (ANNs) as a modeling tool and genetic algorithms (GAs) as an optimizing method for input variables used in ANN modeling was assessed. Hence, as a new technique, it was applied for the prediction and optimization of the plant hormones concentrations and combinations for in vitro proliferation of Garnem (G × N15) rootstock as a case study. Optimizing hormones combination was surveyed by modeling the effects of various concentrations of cytokinin-auxin, i.e., BAP, KIN, TDZ, IBA, and NAA combinations (inputs) on four growth parameters (outputs), i.e., micro-shoots number per explant, length of micro-shoots, developed callus weight (CW) and the quality index (QI) of plantlets. Calculation of statistical values such as R 2 (coefficient of determination) related to the accuracy of ANN-GA models showed a considerably higher prediction accuracy for ANN models, i.e., micro-shoots number: R 2 = 0.81, length of micro-shoots: R 2 = 0.87, CW: R 2 = 0.88, QI: R 2 = 0.87. According to the results, among the input variables, BAP (19.3), KIN (9.64), and IBA (2.63) showed the highest values of variable sensitivity ratio for proliferation rate. The GA showed that media containing 1.02 mg/l BAP in combination with 0.098 mg/l IBA could lead to the optimal proliferation rate (10.53) for G × N15 rootstock. Another objective of the present study was to compare the performance of predicted and optimized cytokinin-auxin combination with the best optimized obtained concentrations of our other experiments. Considering three growth parameters (length of micro-shoots, micro-shoots number, and proliferation rate), the last treatment was found to be superior to the rest of treatments for G × N15 rootstock in vitro multiplication. Very little difference between the ANN predicted and experimental data confirmed high capability of ANN-GA method in predicting new optimized protocols for plant in vitro propagation.

We investigate the problem of dynamic optimal capital growth of a portfolio. A general framework that one strives to maximize the expected logarithm utility of long term growth rate was developed. Exact optimizationalgorithms run into difficulties in this framework and this motivates the investigation of applying simulated annealing optimizedalgorithm to optimize the capital growth of a given portfolio. Empirical results with real financial data indicate that the approach is inspiring for capital growth portfolio.

This paper presents a novel bioinspired algorithm to tackle complex optimization problems: the coral reefs optimization (CRO) algorithm. The CRO algorithm artificially simulates a coral reef, where different corals (namely, solutions to the optimization problem considered) grow and reproduce in coral colonies, fighting by choking out other corals for space in the reef. This fight for space, along with the specific characteristics of the corals' reproduction, produces a robust metaheuristic algorithm shown to be powerful for solving hard optimization problems. In this research the CRO algorithm is tested in several continuous and discrete benchmark problems, as well as in practical application scenarios (i.e., optimum mobile network deployment and off-shore wind farm design). The obtained results confirm the excellent performance of the proposed algorithm and open line of research for further application of the algorithm to real-world problems.

This paper presents a novel bioinspired algorithm to tackle complex optimization problems: the coral reefs optimization (CRO) algorithm. The CRO algorithm artificially simulates a coral reef, where different corals (namely, solutions to the optimization problem considered) grow and reproduce in coral colonies, fighting by choking out other corals for space in the reef. This fight for space, along with the specific characteristics of the corals' reproduction, produces a robust metaheuristic algorithm shown to be powerful for solving hard optimization problems. In this research the CRO algorithm is tested in several continuous and discrete benchmark problems, as well as in practical application scenarios (i.e., optimum mobile network deployment and off-shore wind farm design). The obtained results confirm the excellent performance of the proposed algorithm and open line of research for further application of the algorithm to real-world problems. PMID:25147860

Highlights: ► Defect-free friction stir welds have been produced for AA5083-O/AA7075-O. ► Back-propagation was sufficient for predicting hardness and tensile strength. ► A hybrid multi-objective algorithm is proposed to deal with this MOP. ► Multi-objective particle swarm optimization was used to find the Pareto solutions. ► TOPSIS is used to rank the given alternatives of the Pareto solutions. -- Abstract: Friction Stir Welding (FSW) has been successfully used to weld similar and dissimilar cast and wrought aluminium alloys, especially for aircraft aluminium alloys, that generally present with low weldability by the traditional fusion welding process. This paper focuses on the microstructural and mechanical properties of the Friction Stir Welding (FSW) of AA7075-O to AA5083-O aluminium alloys. Weld microstructures, hardness and tensile properties were evaluated in as-welded condition. Tensile tests indicated that mechanical properties of the joint were better than in the base metals. An Artificial Neural Network (ANN) model was developed to simulate the correlation between the Friction Stir Welding parameters and mechanical properties. Performance of the ANN model was excellent and the model was employed to predict the ultimate tensile strength and hardness of butt joint of AA7075–AA5083 as functions of weld and rotational speeds. The multi-objective particle swarm optimization was used to obtain the Pareto-optimal set. Finally, the Technique for Order Preference by Similarity to the Ideal Solution (TOPSIS) was applied to determine the best compromised solution.

Full Text Available The increasingly complex and rapid transmission dynamics of many infectious diseases necessitates the use of new, more advanced methods for surveillance, early detection, and decision-making. Here, we demonstrate that a new method for optimizing surveillance networks can improve the quality of epidemiological information produced by typical provider-based networks. Using past surveillance and Internet search data, it determines the precise locations where providers should be enrolled. When applied to redesigning the provider-based, influenza-like-illness surveillance network (ILINet for the state of Texas, the method identifies networks that are expected to significantly outperform the existing network with far fewer providers. This optimizednetwork avoids informational redundancies and is thereby more effective than networks designed by conventional methods and a recently published algorithm based on maximizing population coverage. We show further that Google Flu Trends data, when incorporated into a network as a virtual provider, can enhance but not replace traditional surveillance methods.

Full Text Available This paper presents an original and efficient PSO algorithm, which is divided into three phases: (1 stabilization, (2 breadth-first search, and (3 depth-first search. The proposed algorithm, called PSO-3P, was tested with 47 benchmark continuous unconstrained optimization problems, on a total of 82 instances. The numerical results show that the proposed algorithm is able to reach the global optimum. This work mainly focuses on unconstrained optimization problems from 2 to 1,000 variables.

The study deals with the problem of city electrical networksoptimization in big towns and megacities to increase electrical energy quality and decrease real and active power losses in the networks as well as in domestic consumers. The optimization is carried out according to the location selection and separate reactive power source in 10 kW networks of Swarm Intelligence algorithms, in particular, of Particle Swarm one. The problem solution based on Particle Swarm algorithm is determined by variables being discrete quantities and, in addition, there are several local minimums (troughs) to be available for a global minimum to be found. It is proved that the city power supply system optimization is carried out by the additional reactive power source to be installed at consumers location reducing reactive power flow, thereby, ensuring increase of power supply system quality and decrease of power losses in city networks.

Full Text Available Because the increasingly development and use of wireless networks and mobile technologies, was implemented the idea that users of mobile terminals must have access in different wireless networks simultaneously. Therefore one of the main interest points of Next Generation Wireless Networks (NGWNs, refers to the ability to support wireless network access equipment to ensure a high rate of services between different wireless networks. To solve these problems it was necessary to have decision algorithms to decide for each user of mobile terminal, which is the best network at some point, for a service or a specific application that the user needs. Therefore to make these things, different algorithms use the vertical handoff technique. Below are presented a series of algorithms based on vertical handoff technique with a classification of the different existing vertical handoff decision strategies, which tries to solve these issues of wireless network selection at a given time for a specific application of an user. Based on our synthesis on vertical handoff decision strategies given below, we build our strategy based on solutions presented below, taking the most interesting aspect of each one.

Full Text Available The present work was aimed at assessing the possible use of ripe plantain fruit peel as a green-base catalyst in synthesizing Azadirachta indica oil methyl esters (AIOME. The free fatty acid content of the oil (5.81 wt % was initially reduced to 0.90 wt % using methanol: oil at 2.19 v/v, Fe2(SO43 at 6 wt %, time of 15 min and temperature of 65 °C. The pretreated oil was converted to AIOME in a transesterification process with calcined ripe plantain peel ash (CRPPA at 700 °C as catalyst. The process was modeled by artificial neural network and optimized using genetic algorithm. The effectiveness of the developed CRPPA is ascribable to its high K content and microstructural transformation. The reliability of the model obtained was confirmed with a high coefficient of determination (R2 of 0.996 and a low mean relative percentage deviation (MRPD of 8.10%. The best operating variables combination for the process was methanol:oil of 0.73 v/v, CRPPA of 0.65 wt % and time of 57 min while the temperature was kept constant at 65 °C with a corresponding AIOME yield of 99.2 wt %. The results of this work demonstrated the potentials of ripe plantain peels and neem oil as cheap feedstocks for biodiesel production.

In consideration of capacity constraints, fuzzy defect ratio, and fuzzy transport loss ratio, this paper attempted to establish an optimized decision model for production planning and distribution of a multiphase, multiproduct reverse supply chain, which addresses defects returned to original manufacturers, and in addition, develops hybrid algorithms such as Particle Swarm Optimization-Genetic Algorithm (PSO-GA), Genetic Algorithm-Simulated Annealing (GA-SA), and Particle Swarm Optimization-Simulated Annealing (PSO-SA) for solving the optimized model. During a case study of a multi-phase, multi-product reverse supply chain network, this paper explained the suitability of the optimized decision model and the applicability of the algorithms. Finally, the hybrid algorithms showed excellent solving capability when compared with original GA and PSO methods. PMID:24892057

In this paper, a novel nature-inspired optimizationalgorithm has been employed and the trained behaviour of dogs in detecting smell trails is adapted into computational agents for problem solving. The algorithm involves creation of a surface with smell trails and subsequent iteration of the agents in resolving a path. This algorithm can be applied in different computational constraints that incorporate path-based problems. Implementation of the algorithm can be treated as a shortest path problem for a variety of datasets. The simulated agents have been used to evolve the shortest path between two nodes in a graph. This algorithm is useful to solve NP-hard problems that are related to path discovery. This algorithm is also useful to solve many practical optimization problems. The extensive derivation of the algorithm can be enabled to solve shortest path problems.

Full Text Available The development of hybrid algorithms for solving complex optimization problems focuses on enhancing the strengths and compensating for the weakness of two or more complementary approaches. The goal is to intelligently combine the key elements of these approaches to find superior solutions to solve optimization problems. Optimal routing in communication network is considering a complex optimization problem. In this paper we propose a hybrid Hopfield Neural Network (HNN and Tabu Search (TS algorithm, this algorithm called hybrid HNN-TS algorithm. The paradigm of this hybridization is embedded. We embed the short-term memory and tabu restriction features from TS algorithm in the HNN model. The short-term memory and tabu restriction control the neuron selection process in the HNN model in order to get around the local minima problem and find an optimal solution using the HNN model to solve complex optimization problem. The proposed algorithm is intended to find the optimal path for packet transmission in the network which is fills in the field of routing problem. The optimal path that will be selected is depending on 4-tuples (delay, cost, reliability and capacity. Test results show that the propose algorithm can find path with optimal cost and a reasonable number of iterations. It also shows that the complexity of the network model won’t be a problem since the neuron selection is done heuristically.

This work focuses on finding optimal barrier policy for an insurance risk model when the dividends are paid to the share holders according to a barrier strategy. A new approach based on stochastic optimization methods is developed. Compared with the existing results in the literature, more general surplus processes are considered. Precise models of the surplus need not be known; only noise-corrupted observations of the dividends are used. Using barrier-type strategies, a class of stochastic optimizationalgorithms are developed. Convergence of the algorithm is analyzed; rate of convergence is also provided. Numerical results are reported to demonstrate the performance of the algorithm.

In this book, a novel optimization method inspired by a paradigm from nature is introduced. The chemical reactions are used as a paradigm to propose an optimization method that simulates these natural processes. The proposed algorithm is described in detail and then a set of typical complex benchmark functions is used to evaluate the performance of the algorithm. Simulation results show that the proposed optimizationalgorithm can outperform other methods in a set of benchmark functions. This chemical reaction optimization paradigm is also applied to solve the tracking problem for the dynamic model of a unicycle mobile robot by integrating a kinematic and a torque controller based on fuzzy logic theory. Computer simulations are presented confirming that this optimization paradigm is able to outperform other optimization techniques applied to this particular robot application

Mobile sensor networks are considered to consist of a network of mobile robots, each of which has computation, communication and sensing capabilities. Energy efficiency is a critical issue in mobile sensor networks, especially when mobility (i.e., locomotion control), routing (i.e., communications) and sensing are unique characteristics of mobile robots for energy optimization. This thesis focuses on the problem of energy optimization of mobile robotic sensor networks, and the research results can be extended to energy optimization of a network of mobile robots that monitors the environment, or a team of mobile robots that transports materials from stations to stations in a manufacturing environment. On the energy optimization of mobile robotic sensor networks, our research focuses on the investigation and development of distributed optimizationalgorithms to exploit the mobility of robotic sensor nodes for network lifetime maximization. In particular, the thesis studies these five problems: 1. Network-lifetime maximization by controlling positions of networked mobile sensor robots based on local information with distributed optimizationalgorithms; 2. Lifetime maximization of mobile sensor networks with energy harvesting modules; 3. Lifetime maximization using joint design of mobility and routing; 4. Optimal control for network energy minimization; 5. Network lifetime maximization in mobile visual sensor networks. In addressing the first problem, we consider only the mobility strategies of the robotic relay nodes in a mobile sensor network in order to maximize its network lifetime. By using variable substitutions, the original problem is converted into a convex problem, and a variant of the sub-gradient method for saddle-point computation is developed for solving this problem. An optimal solution is obtained by the method. Computer simulations show that mobility of robotic sensors can significantly prolong the lifetime of the whole robotic sensor network while

The basic aims of this work are: to develop reliable algorithms for solving optimization problems involving large stoi- chiometric matrices; to investigate cyclic dependency between metabolic and macromolecular biosynthetic networks; and to quantify the significance of thermodynamic constraints on prokaryotic metabolism.

This book explores fundamental aspects of geometric network optimisation with applications to a variety of real world problems. It presents, for the first time in the literature, a cohesive mathematical framework within which the properties of such optimal interconnection networks can be understood across a wide range of metrics and cost functions. The book makes use of this mathematical theory to develop efficient algorithms for constructing such networks, with an emphasis on exact solutions. Marcus Brazil and Martin Zachariasen focus principally on the geometric structure of optimal interconnection networks, also known as Steiner trees, in the plane. They show readers how an understanding of this structure can lead to practical exact algorithms for constructing such trees. The book also details numerous breakthroughs in this area over the past 20 years, features clearly written proofs, and is supported by 135 colour and 15 black and white figures. It will help graduate students, working mathematicians, ...

ACADEMIC TRAINING LECTURE REGULAR PROGRAMME 1, 2, 3 and 4 June From 11:00 hrs to 12:00 hrs - Main Auditorium bldg. 500 Evolutionary Heuristic Optimization: Genetic Algorithms and Estimation of Distribution Algorithms V. Robles Forcada and M. Perez Hernandez / Univ. de Madrid, Spain In the real world, there exist a huge number of problems that require getting an optimum or near-to-optimum solution. Optimization can be used to solve a lot of different problems such as network design, sets and partitions, storage and retrieval or scheduling. On the other hand, in nature, there exist many processes that seek a stable state. These processes can be seen as natural optimization processes. Over the last 30 years several attempts have been made to develop optimizationalgorithms, which simulate these natural optimization processes. These attempts have resulted in methods such as Simulated Annealing, based on natural annealing processes or Evolutionary Computation, based on biological evolution processes. Geneti...

ENSEIGNEMENT ACADEMIQUE ACADEMIC TRAINING Françoise Benz 73127 academic.training@cern.ch ACADEMIC TRAINING LECTURE REGULAR PROGRAMME 1, 2, 3 and 4 June From 11:00 hrs to 12:00 hrs - Main Auditorium bldg. 500 Evolutionary Heuristic Optimization: Genetic Algorithms and Estimation of Distribution Algorithms V. Robles Forcada and M. Perez Hernandez / Univ. de Madrid, Spain In the real world, there exist a huge number of problems that require getting an optimum or near-to-optimum solution. Optimization can be used to solve a lot of different problems such as network design, sets and partitions, storage and retrieval or scheduling. On the other hand, in nature, there exist many processes that seek a stable state. These processes can be seen as natural optimization processes. Over the last 30 years several attempts have been made to develop optimizationalgorithms, which simulate these natural optimization processes. These attempts have resulted in methods such as Simulated Annealing, based on nat...

To iteratively solve large scale optimization problems in various contexts like planning, operations, design etc., we need to generate descent directions that are based on linear system solutions. Irrespective of the optimizationalgorithm or the solution method employed for the linear systems, ill conditioning introduced by problem characteristics or the algorithm or both need to be addressed. In [GL01] we used an intuitive heuristic approach in scaling linear systems that improved performan...

The widely used backpropagation algorithm for training neural networks based on the gradient descent has a significant drawback of slow convergence. A Gauss-Newton method based recursive least squares (RLS) type algorithm with dynamic error backpropagation is presented to speed-up the learning procedure of neural networks with local recurrent terms. Finally, simulation examples concerning the applications of the RLS type algorithm to identification of nonlinear processes using a local recurrent neural network are also included in this paper

Information Dynamics in Networks: Models and Algorithms In this project, we investigated how network structure interplays with higher level processes in...Models and Algorithms Report Title In this project, we investigated how network structure interplays with higher level processes in online social...Received Paper 1.00 2.00 3.00 . A Note on Modeling Retweet Cascades on Twitter, Workshop on Algorithms and Models for the Web Graph. 09-DEC-15

We present a distributed multi-scale dead-reckoning (MSDR-D) algorithm for network localization that utilizes local distance and angular information for nearby sensors. The algorithm is anchor-free and does not require particular network topology, rigidity of the underlying communication graph, or high average connectivity. The algorithm scales well to large and sparse networks with complex topologies and outperforms previous algorithms when the noise levels are high. The algorithm is simple to implement and is available, along with source code, executables, and experimental results, at http://msdr-d.cs.arizona.edu/.

The typical behavior of optimal solutions to portfolio optimization problems with absolute deviation and expected shortfall models using replica analysis was pioneeringly estimated by S. Ciliberti et al. [Eur. Phys. B. 57, 175 (2007)]; however, they have not yet developed an approximate derivation method for finding the optimal portfolio with respect to a given return set. In this study, an approximation algorithm based on belief propagation for the portfolio optimization problem is presented using the Bethe free energy formalism, and the consistency of the numerical experimental results of the proposed algorithm with those of replica analysis is confirmed. Furthermore, the conjecture of H. Konno and H. Yamazaki, that the optimal solutions with the absolute deviation model and with the mean-variance model have the same typical behavior, is verified using replica analysis and the belief propagation algorithm.

Tuberculosis (TB), one of the most common infectious diseases, requires treatment with multiple antibiotics taken over at least 6 months. This long treatment often results in poor patient-adherence, which can lead to the emergence of multi-drug resistant TB. New antibiotic treatment strategies are sorely needed. New antibiotics are being developed or repurposed to treat TB, but as there are numerous potential antibiotics, dosing sizes and potential schedules, the regimen design space for new treatments is too large to search exhaustively. Here we propose a method that combines an agent-based multi-scale model capturing TB granuloma formation with algorithms for mathematical optimization to identify optimal TB treatment regimens. We define two different single-antibiotic treatments to compare the efficiency and accuracy in predicting optimal treatment regimens of two optimizationalgorithms: genetic algorithms (GA) and surrogate-assisted optimization through radial basis function (RBF) networks. We also illustrate the use of RBF networks to optimize double-antibiotic treatments. We found that while GAs can locate optimal treatment regimens more accurately, RBF networks provide a more practical strategy to TB treatment optimization with fewer simulations, and successfully estimated optimal double-antibiotic treatment regimens. Our results indicate surrogate-assisted optimization can locate optimal TB treatment regimens from a larger set of antibiotics, doses and schedules, and could be applied to solve optimization problems in other areas of research using systems biology approaches. Our findings have important implications for the treatment of diseases like TB that have lengthy protocols or for any disease that requires multiple drugs.

A new algorithm for constructing optimal dyadic decision trees was recently introduced, analyzed, and shown to be very effective for low dimensional data sets. This paper enhances and extends this algorithm by: introducing an adaptive grid search for the regularization parameter that guarantees optimal solutions for all relevant trees sizes, revising the core tree-building algorithm so that its run time is substantially smaller for most regularization parameter values on the grid, and incorporating new data structures and data pre-processing steps that provide significant run time enhancement in practice.

-order derivatives between the mapped coarse model and the fine model at the current iteration point. We also consider an enhanced version in which the input SM coefficients are frequency dependent. The performance of our new algorithms is comparable with the recently published SMIS algorithm when applied......A simple, efficient optimizationalgorithm based on space mapping (SM) is presented. It utilizes input SM to reduce the misalignment between the coarse and fine models of the optimized object over a region of interest, and output space mapping (OSM) to ensure matching of response and first...

We developed a general algorithm for online optimization of accelerator performance, i.e., online tuning, using the performance measure as the objective function. This method, named robust conjugate direction search (RCDS), combines the conjugate direction set approach of Powell's method with a robust line optimizer which considers the random noise in bracketing the minimum and uses parabolic fit of data points that uniformly sample the bracketed zone. Moreover, it is much more robust against noise than traditional algorithms and is therefore suitable for online application. Simulation and experimental studies have been carried out to demonstrate the strength of the new algorithm.

This paper proposes an optimizationalgorithm, the dimension-down iterative algorithm (DDIA), for solving a mixed transportation network design problem (MNDP), which is generally expressed as a mathematical programming with equilibrium constraint (MPEC). The upper level of the MNDP aims to optimize the network performance via both the expansion of the existing links and the addition of new candidate links, whereas the lower level is a traditional Wardrop user equilibrium (UE) problem. The idea of the proposed solution algorithm (DDIA) is to reduce the dimensions of the problem. A group of variables (discrete/continuous) is fixed to optimize another group of variables (continuous/discrete) alternately; then, the problem is transformed into solving a series of CNDPs (continuous network design problems) and DNDPs (discrete network design problems) repeatedly until the problem converges to the optimal solution. The advantage of the proposed algorithm is that its solution process is very simple and easy to apply. Numerical examples show that for the MNDP without budget constraint, the optimal solution can be found within a few iterations with DDIA. For the MNDP with budget constraint, however, the result depends on the selection of initial values, which leads to different optimal solutions (i.e., different local optimal solutions). Some thoughts are given on how to derive meaningful initial values, such as by considering the budgets of new and reconstruction projects separately.

Full Text Available Since few years ago, the parallel processing has been embedded in personal computers by including co-processing units as the graphics processing units resulting in a heterogeneous platform. This paper presents the implementation of swarm algorithms on this platform to solve several functions from optimization problems, where they highlight their inherent parallel processing and distributed control features. In the swarm algorithms, each individual and dimension problem are parallelized by the granularity of the processing system which also offer low communication latency between individuals through the embedded processing. To evaluate the potential of swarm algorithms on graphics processing units we have implemented two of them: the particle swarm optimizationalgorithm and the bacterial foraging optimizationalgorithm. The algorithms’ performance is measured using the acceleration where they are contrasted between a typical sequential processing platform and the NVIDIA GeForce GTX480 heterogeneous platform; the results show that the particle swarm algorithm obtained up to 36.82x and the bacterial foraging swarm algorithm obtained up to 9.26x. Finally, the effect to increase the size of the population is evaluated where we show both the dispersion and the quality of the solutions are decreased despite of high acceleration performance since the initial distribution of the individuals can converge to local optimal solution.

Evolutionary algorithms are widely used in single and multi-objective optimization. They are easy to use and provide solution(s) in one simulation run. They are used in food processing industries for decision making. Food processing presents constrained and unconstrained optimization problems. This paper reviews the ...

In the literature, transport networks are usually treated as homogeneous networks, that is, every node has the same function, simultaneously providing and requiring resources. However, some real networks, such as power grids and supply chain networks, show a far different scenario in which nodes are classified into two categories: supply nodes provide some kinds of services, while demand nodes require them. In this paper, we propose a general transport model for these supply-demand networks, associated with a criterion to quantify their transport capacities. In a supply-demand network with heterogeneous degree distribution, its transport capacity strongly depends on the locations of supply nodes. We therefore design a simulated annealing algorithm to find the near optimal configuration of supply nodes, which remarkably enhances the transport capacity compared with a random configuration and outperforms the degree target algorithm, the betweenness target algorithm, and the greedy method. This work provides a start point for systematically analyzing and optimizing transport dynamics on supply-demand networks.

Full Text Available A direct search algorithm is proposed for minimizing an arbitrary real valued function. The algorithm uses a new function transformation and three simplex-based operations. The function transformation provides global exploration features, while the simplex-based operations guarantees the termination of the algorithm and provides global convergence to a stationary point if the cost function is differentiable and its gradient is Lipschitz continuous. The algorithm’s performance has been extensively tested using benchmark functions and compared to some well-known global optimizationalgorithms. The results of the computational study show that the algorithm combines both simplicity and efficiency and is competitive with the heuristics-based strategies presently used for global optimization.

In this work the results of a genetic algorithm (AG) and a neural recurrent multi state network (RNRME) for optimizing the fuel reload of 5 cycles of the Laguna Verde nuclear power plant (CNLV) are presented. The fuel reload obtained by both methods are compared and it was observed that the RNRME creates better fuel distributions that the AG. Moreover a comparison of the utility for using one or another one techniques is make. (Author)

This textbook covers the fundamentals of optimization, including linear, mixed-integer linear, nonlinear, and dynamic optimization techniques, with a clear engineering focus. It carefully describes classical optimization models and algorithms using an engineering problem-solving perspective, and emphasizes modeling issues using many real-world examples related to a variety of application areas. Providing an appropriate blend of practical applications and optimization theory makes the text useful to both practitioners and students, and gives the reader a good sense of the power of optimization and the potential difficulties in applying optimization to modeling real-world systems. The book is intended for undergraduate and graduate-level teaching in industrial engineering and other engineering specialties. It is also of use to industry practitioners, due to the inclusion of real-world applications, opening the door to advanced courses on both modeling and algorithm development within the industrial engineering ...

Electricite de France (EDF) operates 58 nuclear power plants (NPP), of the Pressurized Water Reactor type. The loading pattern optimization of these NPP is currently done by EDF expert engineers. Within this framework, EDF R and D has developed automatic optimization tools that assist the experts. LOOP is an industrial tool, developed by EDF R and D and based on a simulated annealing algorithm. In order to improve the results of such automatic tools, new optimization methods have to be tested. Ant Colony Optimization (ACO) algorithms are recent methods that have given very good results on combinatorial optimization problems. In order to evaluate the performance of such methods on loading pattern optimization, direct comparisons between LOOP and a mock-up based on the Max-Min Ant System algorithm (a particular variant of ACO algorithms) were made on realistic test-cases. It is shown that the results obtained by the ACO mock-up are very similar to those of LOOP. Future research will consist in improving these encouraging results by using parallelization and by hybridizing the ACO algorithm with local search procedures. (author)

Binary amplitude modulation promises to allow rapid focusing through strongly scattering media with a large number of segments due to the faster update rates of digital micromirror devices (DMDs) compared to spatial light modulators (SLMs). While binary amplitude modulation has a lower theoretical enhancement than phase modulation, the faster update rate should more than compensate for the difference - a factor of π2 /2. Here we present two new algorithms, a genetic algorithm and a transmission matrix algorithm, for optimizing the focus with binary amplitude modulation that achieve enhancements close to the theoretical maximum. Genetic algorithms have been shown to work well in noisy environments and we show that the genetic algorithm performs better than a stepwise algorithm. Transmission matrix algorithms allow complete characterization and control of the medium but require phase control either at the input or output. Here we introduce a transmission matrix algorithm that works with only binary amplitude control and intensity measurements. We apply these algorithms to binary amplitude modulation using a Texas Instruments Digital Micromirror Device. Here we report an enhancement of 152 with 1536 segments (9.90%×N) using a genetic algorithm with binary amplitude modulation and an enhancement of 136 with 1536 segments (8.9%×N) using an intensity-only transmission matrix algorithm.

Full Text Available The paper presents an (ACSA Ant colony search Algorithm for Optimal Reactive Power Optimization and voltage control of power systems. ACSA is a new co-operative agents’ approach, which is inspired by the observation of the behavior of real ant colonies on the topic of ant trial formation and foraging methods. Hence, in the ACSA a set of co-operative agents called "Ants" co-operates to find good solution for Reactive Power Optimization problem. The ACSA is applied for optimal reactive power optimization is evaluated on standard IEEE, 30, 57, 191 (practical test bus system. The proposed approach is tested and compared to genetic algorithm (GA, Adaptive Genetic Algorithm (AGA.

Full Text Available Multicast routing (MR is a technology for delivering network data from some source node(s to a group of destination nodes. The objective of the minimum cost MR (MCMR problem is to find an optimal multicast tree with the minimum cost for MR. This problem is NP complete. In order to tackle the problem, this paper proposes a novel algorithm termed the minimum cost multicast routing ant colony optimization (MCMRACO. Based on the ant colony optimization (ACO framework, the artificial ants in the proposed algorithm use a probabilistic greedy realization of Prim’s algorithm to construct multicast trees. Moving in a cost complete graph (CCG of the network topology, the ants build solutions according to the heuristic and pheromone information. The heuristic information represents problem-specific knowledge for the ants to construct solutions. The pheromone update mechanisms coordinate the ants’ activities by modulating the pheromones. The algorithm can quickly respond to the changes of multicast nodes in a dynamic MR environment. The performance of the proposed algorithm has been compared with published results available in the literature. Results show that the proposed algorithm performs well in both static and dynamic MCMR problems.

In this paper, we propose a novel algorithm-parallel adaptive quantum genetic algorithm-which can rapidly determine the minimum control nodes of arbitrary networks with both control nodes and state nodes. The corresponding network can be fully controlled with the obtained control scheme. We transformed the network controllability issue into a combinational optimization problem based on the Popov-Belevitch-Hautus rank condition. A set of canonical networks and a list of real-world networks were experimented. Comparison results demonstrated that the algorithm was more ideal to optimize the controllability of networks, especially those larger-size networks. We demonstrated subsequently that there were links between the optimal control nodes and some network statistical characteristics. The proposed algorithm provides an effective approach to improve the controllability optimization of large networks or even extra-large networks with hundreds of thousands nodes.

New algorithms are presented for the solution of optimum tolerance assignment problems. The problems considered are defined mathematically as a worst-case problem (WCP), a fixed tolerance problem (FTP), and a variable tolerance problem (VTP). The basic optimization problem without tolerances...... is denoted the zero tolerance problem (ZTP). For solution of the WCP we suggest application of interval arithmetic and also alternative methods. For solution of the FTP an algorithm is suggested which is conceptually similar to algorithms previously developed by the authors for the ZTP. Finally, the VTP...... is solved by a double-iterative algorithm in which the inner iteration is performed by the FTP- algorithm. The application of the algorithm is demonstrated by means of relatively simple numerical examples. Basic properties, such as convergence properties, are displayed based on the examples....

Full Text Available This paper proposes a swarm intelligence algorithm, called firefly mating algorithm (FMA, for solving continuous optimization problems. FMA uses genetic algorithm as the core of the algorithm. The main feature of the algorithm is a novel mating pair selection method which is inspired by the following 2 mating behaviors of fireflies in nature: (i the mutual attraction between males and females causes them to mate and (ii fireflies of both sexes are of the multiple-mating type, mating with multiple opposite sex partners. A female continues mating until her spermatheca becomes full, and, in the same vein, a male can provide sperms for several females until his sperm reservoir is depleted. This new feature enhances the global convergence capability of the algorithm. The performance of FMA was tested with 20 benchmark functions (sixteen 30-dimensional functions and four 2-dimensional ones against FA, ALC-PSO, COA, MCPSO, LWGSODE, MPSODDS, DFOA, SHPSOS, LSA, MPDPGA, DE, and GABC algorithms. The experimental results showed that the success rates of our proposed algorithm with these functions were higher than those of other algorithms and the proposed algorithm also required fewer numbers of iterations to reach the global optima.

Recently, a new metaheuristic called chemical reaction optimization was proposed. This search algorithm, inspired by chemical reactions launched during collisions, inherits several features from other metaheuristics such as simulated annealing and particle swarm optimization. This fact has made it, nowadays, one of the most powerful search algorithms in solving mono-objective optimization problems. In this paper, we propose a multiobjective variant of chemical reaction optimization, called nondominated sorting chemical reaction optimization, in an attempt to exploit chemical reaction optimization features in tackling problems involving multiple conflicting criteria. Since our approach is based on nondominated sorting, one of the main contributions of this paper is the proposal of a new quasi-linear average time complexity quick nondominated sorting algorithm; thereby making our multiobjective algorithm efficient from a computational cost viewpoint. The experimental comparisons against several other multiobjective algorithms on a variety of benchmark problems involving various difficulties show the effectiveness and the efficiency of this multiobjective version in providing a well-converged and well-diversified approximation of the Pareto front.

It is challenging to obtain reliable and optimal mapping between networks for alignment algorithms when both nodal and topological structures are taken into consideration due to the underlying NP-hard problem. Here, we introduce an adaptive hybrid algorithm that combines the classical Hungarian algorithm and the Greedy algorithm (HGA) for the global alignment of biomolecular networks. With this hybrid algorithm, every pair of nodes with one in each network is first aligned based on node information (e.g., their sequence attributes) and then followed by an adaptive and convergent iteration procedure for aligning the topological connections in the networks. For four well-studied protein interaction networks, i.e., C.elegans, yeast, D.melanogaster, and human, applications of HGA lead to improved alignments in acceptable running time. The mapping between yeast and human PINs obtained by the new algorithm has the largest value of common gene ontology (GO) terms compared to those obtained by other existing algorithms, while it still has lower Mean normalized entropy (MNE) and good performances on several other measures. Overall, the adaptive HGA is effective and capable of providing good mappings between aligned networks in which the biological properties of both the nodes and the connections are important.

Full Text Available Gravitational Search Algorithm (GSA is a widely used metaheuristic algorithm. Although fewer parameters in GSA were adjusted, GSA has a slow convergence rate. In this paper, we change the constant acceleration coefficients to be the exponential function on the basis of combination of GSA and PSO (PSO-GSA and propose an improved PSO-GSA algorithm (written as I-PSO-GSA for solving two kinds of classifications: surface water quality and the moving direction of robots. I-PSO-GSA is employed to optimize weights and biases of backpropagation (BP neural network. The experimental results show that, being compared with combination of PSO and GSA (PSO-GSA, single PSO, and single GSA for optimizing the parameters of BP neural network, I-PSO-GSA outperforms PSO-GSA, PSO, and GSA and has better classification accuracy for these two actual problems.

Full Text Available A new image filtering algorithm is proposed. GA-BPN algorithm uses genetic algorithm (GA to decide weights in a back propagation neural network (BPN. It has better global optimal characteristics than traditional optimalalgorithm. In this paper, we used GA-BPN to do image noise filter researching work. Firstly, this paper uses training samples to train GA-BPN as the noise detector. Then, we utilize the well-trained GA-BPN to recognize noise pixels in target image. And at last, an adaptive weighted average algorithm is used to recover noise pixels recognized by GA-BPN. Experiment data shows that this algorithm has better performance than other filters.

Full Text Available Wireless sensor networks are increasingly used for tracking and monitoring dynamic phenomena in urban and natural areas. Spatial coverage is an important issue in sensor networks in order to fulfill the needs of sensing applications. Optimization methods are widely used to efficiently distribute sensor nodes in the network to achieve a desired level of coverage. Most of the existing algorithms do not consider the characteristics of the real environment in the optimization process. In this paper, we propose the integration of contextual information in optimizationalgorithms to improve sensor network coverage. First, we investigate the implication of contextual information in sensor networks. Then, a conceptual framework for local context-aware sensor network deployment optimization method is introduced and related algorithms are presented in detail. Finally, several experiments are carried out to evaluate the validity of the proposed method. The results obtained from these experiments show the effectiveness of our approach in different contextual situations.

An efficient generalized hybrid constructive (GHC) learning algorithm for multioutput radial basis function (RBF) networks is proposed to obtain a compact network with good generalization capability. By this algorithm, one can train the adjustable parameters and determine the optimalnetwork structure simultaneously. First, an initialization method based on the growing and pruning algorithm is utilized to select the important initial hidden neurons and candidate ones. Then, by introducing a generalized hidden matrix, a structured parameter optimizationalgorithm is presented to train multioutput RBF network with fixed size, which combines Levenberg-Marquardt (LM) algorithm with least-square method together. Beginning from an appropriate number of hidden neurons, new neurons chosen from the candidates are added one by one each time when the training entraps into local minima. By incorporating an improved incremental constructive scheme, the training is built on previous results after adding new neurons such that the GHC learning algorithm avoids a trial-and-error procedure. Furthermore, based on the improved computation for LM training, the memory limitation problem is solved. The computational complexity analysis and experimental results demonstrate that better performance is efficiently achieved by this algorithm.

Written by one of the world's leading groups in the area of Bayesian identification, control, and decision making, this book provides the theoretical and algorithmic basis of optimized probabilistic advising. Starting from abstract ideas and formulations, and culminating in detailed algorithms, the book comprises a unified treatment of an important problem of the design of advisory systems supporting supervisors of complex processes. It introduces the theoretical and algorithmic basis of developed advising, relying on novel and powerful combination black-box modelling by dynamic mixture models

Full Text Available An optimization study was conducted on a centrifugal compressor. Eight design variables were chosen from the control points for the Bezier curves which widely influenced the geometric variation; four design variables were selected to optimize the flow passage between the hub and the shroud, and other four design variables were used to improve the performance of the impeller blade. As an optimizationalgorithm, an artificial neural network (ANN was adopted. Initially, the design of experiments was applied to set up the initial data space of the ANN, which was improved during the optimization process using a genetic algorithm. If a result of the ANN reached a higher level, that result was re-calculated by computational fluid dynamics (CFD and was applied to develop a new ANN. The prediction difference between the ANN and CFD was consequently less than 1% after the 6th generation. Using this optimization technique, the computational time for the optimization was greatly reduced and the accuracy of the optimizationalgorithm was increased. The efficiency was improved by 1.4% without losing the pressure ratio, and Pareto-optimal solutions of the efficiency versus the pressure ratio were obtained through the 21st generation.

The exchange market model has mainly been used to study the inequality problem. Although the human society inequality problem is very important, the exchange market models dynamics until stationary state and its capability of ranking individuals is interesting in itself. This study considers the hypothesis that the exchange market model could be understood as an optimization procedure. We present herein the implications for algorithmicoptimization and also the possibility of a new family of exchange market models

A new family of graphs, entangled networks, with optimal properties in many respects, is introduced. By definition, their topology is such that it optimizes synchronizability for many dynamical processes. These networks are shown to have an extremely homogeneous structure: degree, node distance, betweenness, and loop distributions are all very narrow. Also, they are characterized by a very interwoven (entangled) structure with short average distances, large loops, and no well-defined community structure. This family of nets exhibits an excellent performance with respect to other flow properties such as robustness against errors and attacks, minimal first-passage time of random walks, efficient communication, etc. These remarkable features convert entangled networks in a useful concept, optimal or almost optimal in many senses, and with plenty of potential applications in computer science or neuroscience.

In the frame of optimization studies in CNES launcher directorate structure, thermic and material department, the need of an optimization tool based on metaheuristic and finite element models for composite structural dimensioning was underlined.Indeed, composite structures need complex optimization methodologies in order to be really compared to metallic structures with regard to mass, static strength and stiffness constraints (metallic structures using optimization methods better known).After some bibliography research, the use of a genetic algorithm coupled with design of experiment to generate the initial population was chosen. Academic functions were used to validate the optimization process and then it was applied to an industrial study aiming to optimize an interstage skirt with regard to its mass, stiffness and stability (global buckling).

With the rapid development of the energy networks, various forms of renewable energy resources are absorbed into it. Because of the inherent random behaviour of the renewable resources, introducing them into the energy network will destroy the stability of the grids. It is required to use proper energy storages to reduce the uncertain fluctuation from the renewable energy resources. For a concrete model research, this paper presented an explicit method to give suitable capacities of the energy storages in consideration of the economics of the storage, grid losses and the probabilities of the bus voltages violation, for situations of the winds-power generations injected into the power network. Furthermore, the influence of the correlation between the different winds farms on the optimal storage capacity can also be studied by this method.

The use of Partical Swarm OptimizationAlgorithm in this research is to optimize the feature weights on the Voting Feature Interval 5 algorithm so that we can find the model of using PSO algorithm with VFI 5. Optimization of feature weight on Diabetes or Dyspesia data is considered important because it is very closely related to the livelihood of many people, so if there is any inaccuracy in determining the most dominant feature weight in the data will cause death. Increased accuracy by using PSO Algorithm ie fold 1 from 92.31% to 96.15% increase accuracy of 3.8%, accuracy of fold 2 on Algorithm VFI5 of 92.52% as well as generated on PSO Algorithm means accuracy fixed, then in fold 3 increase accuracy of 85.19% Increased to 96.29% Accuracy increased by 11%. The total accuracy of all three trials increased by 14%. In general the Partical Swarm Optimizationalgorithm has succeeded in increasing the accuracy to several fold, therefore it can be concluded the PSO algorithm is well used in optimizing the VFI5 Classification Algorithm.

This dissertation explores the Renewable Energy Integration Problem, and proposes a Genetic Algorithm embedded with a Monte Carlo simulation to solve large instances of the problem that are impractical to solve via full enumeration. The Renewable Energy Integration Problem is defined as finding the optimum set of components to supply the electric demand to a hybrid microgrid. The components considered are solar panels, wind turbines, diesel generators, electric batteries, connections to the power grid and converters, which can be inverters and/or rectifiers. The methodology developed is explained as well as the combinatorial formulation. In addition, 2 case studies of a single objective optimization version of the problem are presented, in order to minimize cost and to minimize global warming potential (GWP) followed by a multi-objective implementation of the offered methodology, by utilizing a non-sorting Genetic Algorithm embedded with a monte Carlo Simulation. The method is validated by solving a small instance of the problem with known solution via a full enumeration algorithm developed by NREL in their software HOMER. The dissertation concludes that the evolutionary algorithms embedded with Monte Carlo simulation namely modified Genetic Algorithms are an efficient form of solving the problem, by finding approximate solutions in the case of single objective optimization, and by approximating the true Pareto front in the case of multiple objective optimization of the Renewable Energy Integration Problem.

A genetic algorithm was used to optimize the power output of multi-junction solar cells. Solar cell operation was modeled using the Silvaco ATLASTM software. The output of the ATLASTM simulation runs served as the input to the genetic algorithm. The genetic algorithm was run as a diffusing computation on a network of eighteen dual processor nodes. Results showed that the genetic algorithm produced better power output optimizations when compared with the results obtained using the hill cli...

The design of water distribution systems is a large class of combinatorial, nonlinear optimization problems with complex constraints such as conservation of mass and energy equations. Since feasible solutions are often extremely complex, traditional optimization techniques are insufficient. Recently, metaheuristic algorithms have been applied to this class of problems because they are highly efficient. In this article, a recently developed optimizer called the mine blast algorithm (MBA) is considered. The MBA is improved and coupled with the hydraulic simulator EPANET to find the optimal cost design for water distribution systems. The performance of the improved mine blast algorithm (IMBA) is demonstrated using the well-known Hanoi, New York tunnels and Balerma benchmark networks. Optimization results obtained using IMBA are compared to those using MBA and other optimizers in terms of their minimum construction costs and convergence rates. For the complex Balerma network, IMBA offers the cheapest network design compared to other optimizationalgorithms.

Time optimal reachability analysis is a novel model based technique for solving scheduling and planning problems. After modeling them as reachability problems using timed automata, a real-time model checker can compute the fastest trace to the goal states which constitutes a time optimal schedule....... We propose distributed computing to accelerate time optimal reachability analysis. We develop five distributed state exploration algorithms, implement them in \\uppaal enabling it to exploit the compute resources of a dedicated model-checking cluster. We experimentally evaluate the implemented...

A novel metaheuristic for continuous optimization problems, named the virus optimizationalgorithm (VOA), is introduced and investigated. VOA is an iteratively population-based method that imitates the behaviour of viruses attacking a living cell. The number of viruses grows at each replication and is controlled by an immune system (a so-called 'antivirus') to prevent the explosive growth of the virus population. The viruses are divided into two classes (strong and common) to balance the exploitation and exploration effects. The performance of the VOA is validated through a set of eight benchmark functions, which are also subject to rotation and shifting effects to test its robustness. Extensive comparisons were conducted with over 40 well-known metaheuristic algorithms and their variations, such as artificial bee colony, artificial immune system, differential evolution, evolutionary programming, evolutionary strategy, genetic algorithm, harmony search, invasive weed optimization, memetic algorithm, particle swarm optimization and simulated annealing. The results showed that the VOA is a viable solution for continuous optimization.

Full Text Available An important performance concern for wireless sensor networks (WSNs is the total energy dissipated by all the nodes in the network over the course of network lifetime. In this paper, we propose a routing algorithm termed as PCA-guided routing algorithm (PCA-RA by exploring the principal component analysis (PCA approach. Our algorithm remarkably reduces energy consumption and prolongs network lifetime by realizing the objective of minimizing the sum of distances between the nodes and the cluster centers in a WSN network. It is demonstrated that the PCA-RA can be efficiently implemented in WSNs by forming a nearly optimal K-means-like clustering structure. In addition, it can decrease the network load while maintaining the accuracy of the sensor measurements during data aggregating process. We validate the efficacy and efficiency of the proposed algorithm by simulations. Both theoretical analyses and simulation results demonstrate that this algorithm can perform significantly with less energy consumption and thus prolong the system lifetime for the networks.

Full Text Available Neural network models have recently made significant achievements in solving vehicle scheduling problems. Adaptive ant colony algorithm provides a new idea for neural networks to solve complex system problems of multiconstrained network intensive vehicle routing models. The pheromone in the path is changed by adjusting the volatile factors in the operation process adaptively. It effectively overcomes the tendency of the traditional ant colony algorithm to fall easily into the local optimal solution and slow convergence speed to search for the global optimal solution. The multiconstrained network intensive vehicle routing algorithm based on adaptive ant colony algorithm in this paper refers to the interaction between groups. Adaptive transfer and pheromone update strategies are introduced based on the traditional ant colony algorithm to optimize the selection, update, and coordination mechanisms of the algorithm further. Thus, the search task of the objective function for a feasible solution is completed by the search ants. Through the division and collaboration of different kinds of ants, pheromone adaptive strategy is combined with polymorphic ant colony algorithm. It can effectively overcome some disadvantages, such as premature stagnation, and has a theoretical significance to the study of large-scale multiconstrained vehicle routing problems in complex traffic network systems.

Full Text Available This paper takes Bao Steel logistics automated warehouse system as an example. The premise is to maintain the focus of the shelf below half of the height of the shelf. As a result, the cost time of getting or putting goods on the shelf is reduced, and the distance of the same kind of goods is also reduced. Construct a multiobjective optimization model, using genetic algorithm to optimize problem. At last, we get a local optimal solution. Before optimization, the average cost time of getting or putting goods is 4.52996 s, and the average distance of the same kinds of goods is 2.35318 m. After optimization, the average cost time is 4.28859 s, and the average distance is 1.97366 m. After analysis, we can draw the conclusion that this model can improve the efficiency of cargo storage.

This book not only provides a comprehensive introduction to neural-based PCA methods in control science, but also presents many novel PCA algorithms and their extensions and generalizations, e.g., dual purpose, coupled PCA, GED, neural based SVD algorithms, etc. It also discusses in detail various analysis methods for the convergence, stabilizing, self-stabilizing property of algorithms, and introduces the deterministic discrete-time systems method to analyze the convergence of PCA/MCA algorithms. Readers should be familiar with numerical analysis and the fundamentals of statistics, such as the basics of least squares and stochastic algorithms. Although it focuses on neural networks, the book only presents their learning law, which is simply an iterative algorithm. Therefore, no a priori knowledge of neural networks is required. This book will be of interest and serve as a reference source to researchers and students in applied mathematics, statistics, engineering, and other related fields.

This book presents efficient metaheuristic algorithms for optimal design of structures. Many of these algorithms are developed by the author and his colleagues, consisting of Democratic Particle Swarm Optimization, Charged System Search, Magnetic Charged System Search, Field of Forces Optimization, Dolphin Echolocation Optimization, Colliding Bodies Optimization, Ray Optimization. These are presented together with algorithms which were developed by other authors and have been successfully applied to various optimization problems. These consist of Particle Swarm Optimization, Big Bang-Big Crunch Algorithm, Cuckoo Search Optimization, Imperialist Competitive Algorithm, and Chaos Embedded Metaheuristic Algorithms. Finally a multi-objective optimization method is presented to solve large-scale structural problems based on the Charged System Search algorithm. The concepts and algorithms presented in this book are not only applicable to optimization of skeletal structures and finite element models, but can equally ...

This book presents efficient metaheuristic algorithms for optimal design of structures. Many of these algorithms are developed by the author and his colleagues, consisting of Democratic Particle Swarm Optimization, Charged System Search, Magnetic Charged System Search, Field of Forces Optimization, Dolphin Echolocation Optimization, Colliding Bodies Optimization, Ray Optimization. These are presented together with algorithms which were developed by other authors and have been successfully applied to various optimization problems. These consist of Particle Swarm Optimization, Big Bang-Big Crunch Algorithm, Cuckoo Search Optimization, Imperialist Competitive Algorithm, and Chaos Embedded Metaheuristic Algorithms. Finally a multi-objective optimization method is presented to solve large-scale structural problems based on the Charged System Search algorithm. The concepts and algorithms presented in this book are not only applicable to optimization of skeletal structures and finite element models, but can equally ...

Full Text Available The uncertainty and regularity of wind power generation are caused by wind resources’ intermittent and randomness. Such volatility brings severe challenges to the wind power grid. The requirements for ultrashort-term and short-term wind power forecasting with high prediction accuracy of the model used, have great significance for reducing the phenomenon of abandoned wind power , optimizing the conventional power generation plan, adjusting the maintenance schedule and developing real-time monitoring systems. Therefore, accurate forecasting of wind power generation is important in electric load forecasting. The echo state network (ESN is a new recurrent neural network composed of input, hidden layer and output layers. It can approximate well the nonlinear system and achieves great results in nonlinear chaotic time series forecasting. Besides, the ESN is simpler and less computationally demanding than the traditional neural network training, which provides more accurate training results. Aiming at addressing the disadvantages of standard ESN, this paper has made some improvements. Combined with the complementary advantages of particle swarm optimization and tabu search, the generalization of ESN is improved. To verify the validity and applicability of this method, case studies of multitime scale forecasting of wind power output are carried out to reconstruct the chaotic time series of the actual wind power generation data in a certain region to predict wind power generation. Meanwhile, the influence of seasonal factors on wind power is taken into consideration. Compared with the classical ESN and the conventional Back Propagation (BP neural network, the results verify the superiority of the proposed method.

In this paper, we consider a distributed robust optimization (DRO) problem, where multiple agents in a networked system cooperatively minimize a global convex objective function with respect to a global variable under the global constraints. The objective function can be represented by a sum of local objective functions. The global constraints contain some uncertain parameters which are partially known, and can be characterized by some inequality constraints. After problem transformation, we adopt the Lagrangian primal-dual method to solve this problem. We prove that the primal and dual optimal solutions of the problem are restricted in some specific sets, and we give a method to construct these sets. Then, we propose a DRO algorithm to find the primal-dual optimal solutions of the Lagrangian function, which consists of a subgradient step, a projection step, and a diffusion step, and in the projection step of the algorithm, the optimized variables are projected onto the specific sets to guarantee the boundedness of the subgradients. Convergence analysis and numerical simulations verifying the performance of the proposed algorithm are then provided. Further, for nonconvex DRO problem, the corresponding approach and algorithm framework are also provided.

In this paper, the resource allocation in an underlay cognitive radio network is considered, in which the SUs can operate on the same spectrum band with the primary users simultaneously. In order to maximize the total throughput of SUs, an optimal selection matching (OSM) algorithm is proposed to allocate the resource in the underlay cognitive radio network. It makes every SU find the satisfied spectrum resource, and meanwhile the interference from SUs is induced. Simulation results show that the proposed algorithm has the significant improvement in total throughput compared with others.

In recent years, the wind power industry has focused its efforts on solving the Wind Farm Layout Optimization (WFLO) problem. Wind resource assessment is a pivotal step in optimizing the wind-farm design and siting and, in determining whether a project is economically feasible or not. In the present work, three (3) different optimization methods are proposed for the solution of the WFLO: (i) A modified Viral System Algorithm applied to the optimization of the proper location of the components in a wind-farm to maximize the energy output given a stated wind environment of the site. The optimization problem is formulated as the minimization of energy cost per unit produced and applies a penalization for the lack of system reliability. The viral system algorithm utilized in this research solves three (3) well-known problems in the wind-energy literature; (ii) a new multiple objective evolutionary algorithm to obtain optimal placement of wind turbines while considering the power output, cost, and reliability of the system. The algorithm presented is based on evolutionary computation and the objective functions considered are the maximization of power output, the minimization of wind farm cost and the maximization of system reliability. The final solution to this multiple objective problem is presented as a set of Pareto solutions and, (iii) A hybrid viral-based optimizationalgorithm adapted to find the proper component configuration for a wind farm with the introduction of the universal generating function (UGF) analytical approach to discretize the different operating or mechanical levels of the wind turbines in addition to the various wind speed states. The proposed methodology considers the specific probability functions of the wind resource to describe their proper behaviors to account for the stochastic comportment of the renewable energy components, aiming to increase their power output and the reliability of these systems. The developed heuristic considers a

Full Text Available IoT is the technical basis to realize the CPS (Cyber Physical System for distribution networks, with which the complex system becomes more intelligent and controllable. Because of the multihop and self-organization characteristics, the large-scale heterogeneous CPS network becomes more difficult to plan. Using topological potential theory, one of typical big data analysis technologies, this paper proposed a novel optimal CPS planning model. Topological potential equalization is considered as the optimization objective function in heterogeneous CPS network with the constraints of communication requirements, physical infrastructures, and network reliability. An improved binary particle swarm optimizationalgorithm is proposed to solve this complex optimal problem. Two IEEE classic examples are adopted in the simulation, and the results show that, compared with benchmark algorithms, our proposed method can provide an effective topology optimization scheme to improve the network reliability and transmitting performance.

An artificial neural networkalgorithm for continuous minimization is developed and applied to the case of numerical particle loading. It is shown that higher-order moments of the probability distribution function can be efficiently renormalized using this technique. A general neural network for the renormalization of an arbitrary number of moments is given

The University of Lowell and Westinghouse Electric Corporation are involved in a joint effort to evaluate the potential benefits of generalized/depletion perturbation theory (GPT/DTP) methods for a variety of light water reactor (LWR) physics applications. One part of that work has focused on the development of a GPT-based optimizationalgorithm for the overall design, analysis, and optimization of LWR reload cores. The use of GPT sensitivity data in formulating the fuel management optimization problem is conceptually straightforward; it is the actual execution of the concept that is challenging. Thus, the purpose of this paper is to address some of the major difficulties, to outline our approach to these problems, and to present some illustrative examples of an efficient GTP-based optimization scheme

Presenting the concept and design and implementation of configurable intelligent optimizationalgorithms in manufacturing systems, this book provides a new configuration method to optimize manufacturing processes. It provides a comprehensive elaboration of basic intelligent optimizationalgorithms, and demonstrates how their improvement, hybridization and parallelization can be applied to manufacturing. Furthermore, various applications of these intelligent optimizationalgorithms are exemplified in detail, chapter by chapter. The intelligent optimizationalgorithm is not just a single algorit

Full Text Available Transmission Network Expansion Planning (TNEP is a basic part of power network planning that determines where, when and how many new transmission lines should be added to the network. So, the TNEP is an optimization problem in which the expansion purposes are optimized. Artificial Intelligence (AI tools such as Genetic Algorithm (GA, Simulated Annealing (SA, Tabu Search (TS and Artificial Neural Networks (ANNs are methods used for solving the TNEP problem. Today, by using the hybridization models of AI tools, we can solve the TNEP problem for large-scale systems, which shows the effectiveness of utilizing such models. In this paper, a new approach to the hybridization model of Probabilistic Neural Networks (PNNs and Harmony Search Algorithm (HSA was used to solve the TNEP problem. Finally, by considering the uncertain role of the load based on a scenario technique, this proposed model was tested on the Garver’s 6-bus network.

In view of the various orders of FAW (First Automotive Works) International Logistics Co., Ltd., the SLP method is used to optimize the layout of the warehousing units in the enterprise, thus the warehouse logistics is optimized and the external processing speed of the order is improved. In addition, the relevant intelligent algorithms for optimizing the stocking route problem are analyzed. The ant colony algorithm and genetic algorithm which have good applicability are emphatically studied. The parameters of ant colony algorithm are optimized by genetic algorithm, which improves the performance of ant colony algorithm. A typical path optimization problem model is taken as an example to prove the effectiveness of parameter optimization.

Full Text Available The application of metaheuristic algorithms to optimization problems has been very important during the last decades. The main advantage of these techniques is their flexibility and robustness, which allows them to be applied to a wide range of problems. In this work we concentrate on metaheuristics based on Simulated Annealing, Tabu Search and Variable Neighborhood Search trajectory whose main characteristic is that they start from a point and through the exploration of the neighborhood vary the current solution, forming a trajectory. By means of the instances of the selected combinatorial problems, a computational experimentation is carried out that illustrates the behavior of the algorithmic methods to solve them. The main objective of this work is to perform the study and comparison of the results obtained for the selected trajectories metaheuristics in its application for the resolution of a set of academic problems of combinatorial optimization.

The authors have been using genetic algorithms to study the structures of atomic clusters and related problems. This is a problem where local minima are easy to locate, but barriers between the many minima are large, and the number of minima prohibit a systematic search. They use a novel mating algorithm that preserves some of the geometrical relationship between atoms, in order to ensure that the resultant structures are likely to inherit the best features of the parent clusters. Using this approach, they have been able to find lower energy structures than had been previously obtained. Most recently, they have been able to turn around the building block idea, using optimized structures from the GA to learn about systematic structural trends. They believe that an effective GA can help provide such heuristic information, and (conversely) that such information can be introduced back into the algorithm to assist in the search process

Full Text Available We study some nonlinear gossip algorithms for wireless sensor networks. Firstly, two types of nonlinear single gossip algorithms are proposed. By using Lyapunov theory, Lagrange mean value theorem, and stochastic Lasalle’s invariance principle, we prove that the nonlinear single gossip algorithms can converge to the average of initial states with probability one. Secondly, two types of nonlinear multigossip algorithms are also presented and the convergence is proved by the same methods. Finally, computer simulation is also given to show the validity of the theoretical results.

The efficient frontier is a core concept in Modern Portfolio Theory. Based on this idea, we will construct optimal trading curves for different types of portfolios. These curves correspond to the algorithmic trading strategies that minimize the expected transaction costs, i.e. the joint effect of market impact and market risk. We will study five portfolio trading strategies. For the first three (single-asset, general multi-asseet and balanced portfolios) we will assume that the underlyings fo...

A new approach to optimize the parameters of a gradient-based optical flow model using a parallel genetic algorithm (GA) is proposed. The main characteristics of the optical flow algorithm are its bio-inspiration and robustness against contrast, static patterns and noise, besides working consistently with several optical illusions where other algorithms fail. This model depends on many parameters which conform the number of channels, the orientations required, the length and shape of the kernel functions used in the convolution stage, among many more. The GA is used to find a set of parameters which improve the accuracy of the optical flow on inputs where the ground-truth data is available. This set of parameters helps to understand which of them are better suited for each type of inputs and can be used to estimate the parameters of the optical flow algorithm when used with videos that share similar characteristics. The proposed implementation takes into account the embarrassingly parallel nature of the GA and uses the OpenMP Application Programming Interface (API) to speedup the process of estimating an optimal set of parameters. The information obtained in this work can be used to dynamically reconfigure systems, with potential applications in robotics, medical imaging and tracking.

Artificial neural networks (Ann) are widely used; it which consist of an input layer, one or more hidden layers and an output layer; these layers contain neurons and each has connections called weights, where the knowledge are allowed and let to Ann solve problems proposed. These Ann is used to reconstruction of the energy spectrum of neutrons from count rates and develop Bonner sphere neutron dosimetry. Currently, we have developed Ann with high performance and generalization ability. Determine your optimal architecture is usually a difficult task, an exhaustive search of all possible combinations of parameters is rarely possible further training of the neural network with random initial weights can cause two major drawbacks: it can stuck in local minima or converge very slowly. In this project it will be used Genetic Algorithms (Ga); which are based on the principle or analogy of evolution through natural selection and has shown to be very effective in optimizing complex search functions and large spaces or to find a near optimal overall efficiency. The aim is to decrease the architecture in number of hidden neurons and therefore the total number of connections is reducing. The benefits obtained by optimizing the network are that the number of connections would be considerably smaller and thus the computational complexity, hardware integration, resources will be lower such that will allow to be even more viable implemented. To use the Ga three problems must be solve: 1) coding the problem into chromosomes. 2) Construct a fitness function. 3) Proper selection of genetic operators; crossover, selection, mutation. As a result, the scientific knowledge obtained can to be applied to similar problems having a reference parameters used and their impact on the optimization would to be generated. It concluded that the input layer and output are subject to the problem; the Ga propose the optimal number of neurons in the hidden layer without losing the quality of the

In order to optimize the distribution problem of power grid sources, an optimized particle swarm optimizationalgorithm is proposed. First, the concept of multi-objective optimization and the Pareto solution set are enumerated. Then, the performance of the classical genetic algorithm, the classical particle swarm optimizationalgorithm and the improved particle swarm optimizationalgorithm are analyzed. The three algorithms are simulated respectively. Compared with the test results of each algorithm, the superiority of the algorithm in convergence and optimization performance is proved, which lays the foundation for subsequent micro-grid power optimization configuration solution.

Full Text Available As the size of the Web continues to grow, searching it for useful information has become more difficult. Focused crawler intends to explore the Web conform to a specific topic. This paper discusses the problems caused by local searching algorithms. Crawler can be trapped within a limited Web community and overlook suitable Web pages outside its track. A genetic algorithm as a global searching algorithm is modified to address the problems. The genetic algorithm is used to optimize Web crawling and to select more suitable Web pages to be fetched by the crawler. Several evaluation experiments are conducted to examine the effectiveness of the approach. The crawler delivers collections consist of 3396 Web pages from 5390 links which had been visited, or filtering rate of Roulette-Wheel selection at 63% and precision level at 93% in 5 different categories. The result showed that the utilization of genetic algorithm had empowered focused crawler to traverse the Web comprehensively, despite it relatively small collections. Furthermore, it brought up a great potential for building an exemplary collections compared to traditional focused crawling methods.

Full Text Available The need to process and analyze large data volumes, as well as to convey the information contained therein to decision makers naturally led to the development of OLAP systems. Similarly to SGBDs, OLAP systems must ensure optimum access to the storage environment. Although there are several ways to optimize database systems, implementing a correct data indexing solution is the most effective and less costly. Thus, OLAP uses indexing algorithms for relational data and n-dimensional summarized data stored in cubes. Today database systems implement derived indexing algorithms based on well-known Tree, Bitmap and Hash indexing algorithms. This is because no indexing algorithm provides the best performance for any particular situation (type, structure, data volume, application. This paper presents a new n-dimensional cube indexing algorithm, derived from the well known B-Tree index, which indexes data stored in data warehouses taking in consideration their multi-dimensional nature and provides better performance in comparison to the already implemented Tree-like index types.

We study {ital budget constrained optimalnetwork upgrading problems}. Such problems aim at finding optimal strategies for improving a network under some cost measure subject to certain budget constraints. Given an edge weighted graph {ital G(V,E)}, in the {ital edge based upgrading model}, it is assumed that each edge {ital e} of the given network has an associated function {ital c(e)} that specifies for each edge {ital e} the amount by which the length {ital l(e)} is to be reduced. In the {ital node based upgrading model} a node {ital v} can be upgraded at an expense of cost {ital (v)}. Such an upgrade reduces the cost of each edge incident on {ital v} by a fixed factor {rho}, where 0 < {rho} < 1. For a given budget, {ital B}, the goal is to find an improvement strategy such that the total cost of reduction is a most the given budget {ital B} and the cost of a subgraph (e.g. minimum spanning tree) under the modified edge lengths is the best over all possible strategies which obey the budget constraint. Define an ({alpha},{beta})-approximation algorithm as a polynomial-time algorithm that produces a solution within {alpha} times the optimal function value, violating the budget constraint by a factor of at most {Beta}. The results obtained in this paper include the following 1. We show that in general the problem of computing optimal reduction strategy for modifying the network as above is {bold NP}-hard. 2. In the node based model, we show how to devise a near optimal strategy for improving the bottleneck spanning tree. The algorithms have a performance guarantee of (2 ln {ital n}, 1). 3. for the edge based improvement problems we present improved (in terms of performance and time) approximation algorithms. 4. We also present pseudo-polynomial time algorithms (extendible to polynomial time approximation schemes) for a number of edge/node based improvement problems when restricted to the class of treewidth-bounded graphs.

In this paper, we propose a novel algorithm—parallel adaptive quantum genetic algorithm—which can rapidly determine the minimum control nodes of arbitrary networks with both control nodes and state nodes. The corresponding network can be fully controlled with the obtained control scheme. We transformed the network controllability issue into a combinational optimization problem based on the Popov-Belevitch-Hautus rank condition. A set of canonical networks and a list of real-world networks were experimented. Comparison results demonstrated that the algorithm was more ideal to optimize the controllability of networks, especially those larger-size networks. We demonstrated subsequently that there were links between the optimal control nodes and some network statistical characteristics. The proposed algorithm provides an effective approach to improve the controllability optimization of large networks or even extra-large networks with hundreds of thousands nodes. PMID:29554140

Full Text Available In this work, a journal bearing optimization process has been developed and is divided into two stages. Each one has a set of decision variables and custom objectives aggregating performances with a weighting strategy. The performance functions used are an artificial neural network, trained with Reynolds equation solutions, and a CFD simulation of the bearings carried out with commercial software. The results show the capabilities of the algorithm to design and optimize journal bearings by reducing both power loss and mass flow with respect to ones designed with traditional methods, as well as by minimizing the maximum and average temperature.

Full Text Available The back propagation neural network (BPNN algorithm can be used as a supervised classification in the processing of remote sensing image classification. But its defects are obvious: falling into the local minimum value easily, slow convergence speed, and being difficult to determine intermediate hidden layer nodes. Genetic algorithm (GA has the advantages of global optimization and being not easy to fall into local minimum value, but it has the disadvantage of poor local searching capability. This paper uses GA to generate the initial structure of BPNN. Then, the stable, efficient, and fast BP classification network is gotten through making fine adjustments on the improved BP algorithm. Finally, we use the hybrid algorithm to execute classification on remote sensing image and compare it with the improved BP algorithm and traditional maximum likelihood classification (MLC algorithm. Results of experiments show that the hybrid algorithm outperforms improved BP algorithm and MLC algorithm.

The optimization of real scalar functions of quaternion variables, such as the mean square error or array output power, underpins many practical applications. Solutions typically require the calculation of the gradient and Hessian. However, real functions of quaternion variables are essentially nonanalytic, which are prohibitive to the development of quaternion-valued learning systems. To address this issue, we propose new definitions of quaternion gradient and Hessian, based on the novel generalized Hamilton-real (GHR) calculus, thus making a possible efficient derivation of general optimizationalgorithms directly in the quaternion field, rather than using the isomorphism with the real domain, as is current practice. In addition, unlike the existing quaternion gradients, the GHR calculus allows for the product and chain rule, and for a one-to-one correspondence of the novel quaternion gradient and Hessian with their real counterparts. Properties of the quaternion gradient and Hessian relevant to numerical applications are also introduced, opening a new avenue of research in quaternion optimization and greatly simplified the derivations of learning algorithms. The proposed GHR calculus is shown to yield the same generic algorithm forms as the corresponding real- and complex-valued algorithms. Advantages of the proposed framework are illuminated over illustrative simulations in quaternion signal processing and neural networks.

Low cost, high reliability and easy maintenance are key criteria in the design of routing protocols for wireless sensor networks (WSNs). This paper investigates the existing ant colony optimization (ACO)-based WSN routing algorithms and the minimum hop count WSN routing algorithms by reviewing their strengths and weaknesses. We also consider the critical factors of WSNs, such as energy constraint of sensor nodes, network load balancing and dynamic network topology. Then we propose a hybrid routing algorithm that integrates ACO and a minimum hop count scheme. The proposed algorithm is able to find the optimal routing path with minimal total energy consumption and balanced energy consumption on each node. The algorithm has unique superiority in terms of searching for the optimal path, balancing the network load and the network topology maintenance. The WSN model and the proposed algorithm have been implemented using C++. Extensive simulation experimental results have shown that our algorithm outperforms several other WSN routing algorithms on such aspects that include the rate of convergence, the success rate in searching for global optimal solution, and the network lifetime.

Until recently, a lot of work has been done to develop algorithms that utilize the distributed structure of an ad hoc wireless sensor network to estimate a certain parameter of interest. However, all these algorithms assume that the input regressor data is available to the sensors, but this data is not always available to the sensors. In such cases, blind estimation of the required parameter is needed. This work formulates two newly developed blind block-recursive algorithms based on singular value decomposition (SVD) and Cholesky factorization-based techniques. These adaptive algorithms are then used for blind estimation in a wireless sensor network using diffusion of data among cooperative sensors. Simulation results show that the performance greatly improves over the case where no cooperation among sensors is involved.

Inside the aspects to be improved in the service of power distribution, reducing Joule losses (technical losses) is vital. This paper proposes a method of reconfiguration of primary distribution networks for minimal losses through the use of genetic algorithms. The same is applied to a multi-node system with feeders, tending to minimize these losses and maintain the voltage in line with current legislation. As a result of the implementation of this system, we found that the optimization method used is able to find the best solution (the optimal solution) among all possible combinations where the switch operations. Also, check your flexibility to adapt to the restrictions of radiality and voltage level, in less time than that required for an exhaustive search.

A laser is one of the most important experimental tools. In synchrotron radiation field, lasers are widely used for experiments with Pump-Probe techniques. Especially for Xray-FELs, a laser has important roles as a seed light source or photo-cathode-illuminating light source to generate a high brightness electron bunch. The controls of laser pulse characteristics are required for many kinds of experiments. However, the laser should be tuned and customized for each requirement by laser experts. The automatic tuning of laser is required to realize with some sophisticated algorithms. The metaheuristic algorithm is one of the useful candidates to find one of the best solutions as acceptable as possible. The metaheuristic laser tuning system is expected to save our human resources and time for the laser preparations. I have shown successful results on a metaheuristic algorithm based on a genetic algorithm to optimize spatial (transverse) laser profiles and a hill climbing method extended with a fuzzy set theory to choose one of the best laser alignments automatically for each experimental requirement. (author)

In many cases, the topological structures of a complex network are unknown or uncertain, and it is of significance to identify the exact topological structure. An optimization-based method of identifying the topological structure of a complex network is proposed in this paper. Identification of the exact network topological structure is converted into a minimal optimization problem by using the estimated network. Then, an improved quantum-behaved particle swarm optimizationalgorithm is used to solve the optimization problem. Compared with the previous adaptive synchronization-based method, the proposed method is simple and effective and is particularly valid to identify the topological structure of synchronization complex networks. In some cases where the states of a complex network are only partially observable, the exact topological structure of a network can also be identified by using the proposed method. Finally, numerical simulations are provided to show the effectiveness of the proposed method. (general)

In this study, graphite oxide (GO) nano according to Hummers method was synthesized and subsequently was used for the removal of methylene blue (MB) and brilliant green (BG). The detail information about the structure and physicochemical properties of GO are investigated by different techniques such as XRD and FTIR analysis. The influence of solution pH, initial dye concentration, contact time and adsorbent dosage was examined in batch mode and optimum conditions was set as pH=7.0, 2 mg of GO and 10 min contact time. Employment of equilibrium isotherm models for description of adsorption capacities of GO explore the good efficiency of Langmuir model for the best presentation of experimental data with maximum adsorption capacity of 476.19 and 416.67 for MB and BG dyes in single solution. The analysis of adsorption rate at various stirring times shows that both dyes adsorption followed a pseudo second-order kinetic model with cooperation with interparticle diffusion model. Subsequently, the adsorption data as new combination of artificial neural network was modeled to evaluate and obtain the real conditions for fast and efficient removal of dyes. A three-layer artificial neural network (ANN) model is applicable for accurate prediction of dyes removal percentage from aqueous solution by GO following conduction of 336 experimental data. The network was trained using the obtained experimental data at optimum pH with different GO amount (0.002-0.008 g) and 5-40 mg/L of both dyes over contact time of 0.5-30 min. The ANN model was able to predict the removal efficiency with Levenberg-Marquardt algorithm (LMA), a linear transfer function (purelin) at output layer and a tangent sigmoid transfer function (tansig) at hidden layer with 10 and 11 neurons for MB and BG dyes, respectively. The minimum mean squared error (MSE) of 0.0012 and coefficient of determination (R(2)) of 0.982 were found for prediction and modeling of MB removal, while the respective value for BG was the

Many methods are available for finding x*E Rn which minimizes the real value function f(x), some of which are Fibonacci Search Algorithm, Quadratic Search Algorithm, Convergence Algorithm and Cubic Search Algorithm. In this research work, existing algorithms used in single variable optimization problems are critically ...

In recent years, deep learning methods applying unsupervised learning to train deep layers of neural networks have achieved remarkable results in numerous fields. In the past, many genetic algorithms based methods have been successfully applied to training neural networks. In this paper, we extend previous work and propose a GA-assisted method for deep learning. Our experimental results indicate that this GA-assisted approach improves the performance of a deep autoencoder, producing a sparser...

non-linearity, so there is no guarantee to find the optimal solution. Moreover, it has generated a great number of variables. Therefore, a heuristic method is required to find near optimal solution with reasonable computation time. For this reason, a genetic algorithm (GA-based procedure is proposed. The proposed procedure then is applied to the same problem as discussed in the basic model. The results indicated that there is significant improvement on hub locations. Flows are successfully consolidated to several big ports as expected. With regards to spoke allocations, however, spokes are not fairly allocated.Keywords: Hub and Spoke Model; Marine Transportation; Genetic Algorithm

Machine learning has the potential to dramatically accelerate high-throughput approaches to materials design, as demonstrated by successes in biomolecular design and hard materials design. However, in the search for new soft materials exhibiting properties and performance beyond those previously achieved, machine learning approaches are frequently limited by two shortcomings. First, because they are intrinsically interpolative, they are better suited to the optimization of properties within the known range of accessible behavior than to the discovery of new materials with extremal behavior. Second, they require large pre-existing data sets, which are frequently unavailable and prohibitively expensive to produce. Here we describe a new strategy, the neural-network-biased genetic algorithm (NBGA), for combining genetic algorithms, machine learning, and high-throughput computation or experiment to discover materials with extremal properties in the absence of pre-existing data. Within this strategy, predictions from a progressively constructed artificial neural network are employed to bias the evolution of a genetic algorithm, with fitness evaluations performed via direct simulation or experiment. In effect, this strategy gives the evolutionary algorithm the ability to "learn" and draw inferences from its experience to accelerate the evolutionary process. We test this algorithm against several standard optimization problems and polymer design problems and demonstrate that it matches and typically exceeds the efficiency and reproducibility of standard approaches including a direct-evaluation genetic algorithm and a neural-network-evaluated genetic algorithm. The success of this algorithm in a range of test problems indicates that the NBGA provides a robust strategy for employing informatics-accelerated high-throughput methods to accelerate materials design in the absence of pre-existing data.

This paper proposes new particle swarm optimization (PSO) methods that use chaotic maps for parameter adaptation. This has been done by using of chaotic number generators each time a random number is needed by the classical PSO algorithm. Twelve chaos-embedded PSO methods have been proposed and eight chaotic maps have been analyzed in the benchmark functions. It has been detected that coupling emergent results in different areas, like those of PSO and complex dynamics, can improve the quality of results in some optimization problems. It has been also shown that, some of the proposed methods have somewhat increased the solution quality, that is in some cases they improved the global searching capability by escaping the local solutions.

This paper proposes new particle swarm optimization (PSO) methods that use chaotic maps for parameter adaptation. This has been done by using of chaotic number generators each time a random number is needed by the classical PSO algorithm. Twelve chaos-embedded PSO methods have been proposed and eight chaotic maps have been analyzed in the benchmark functions. It has been detected that coupling emergent results in different areas, like those of PSO and complex dynamics, can improve the quality of results in some optimization problems. It has been also shown that, some of the proposed methods have somewhat increased the solution quality, that is in some cases they improved the global searching capability by escaping the local solutions.

Many techniques exist for developing optimal controllers. This paper investigates genetic algorithms as a means of finding optimal solutions over a parameter space. In particular, the genetic algorithm is applied to optimal tuning of a governor for a hydrogenerator plant. Analog and digital simulation methods are compared for use in conjunction with the genetic algorithmoptimization process. It is shown that analog plant simulation provides advantages in speed over digital plant simulation. This speed advantage makes application of the genetic algorithm in an actual plant environment feasible. Furthermore, the genetic algorithm is shown to possess the ability to reject plant noise and other system anomalies in its search for optimizing solutions

Since many realistic networks such as wireless sensor/ad hoc networks usually do not agree very well with the basic network models such as small-word and scale-free models, we often need to obtain some expected structural features such as a small average path length and a regular degree distribution while optimizing the connectivity of these networks. Although a minor addition of links for optimizingnetwork connectivity is not likely to change the structural properties of a network, it is necessary to investigate the impact of link addition on network properties as the number of the added links increases. However, to the best of our knowledge, the study of that problem has not been found so far. Furthermore, two closely related questions to that problem, i.e., how to measure and how to improve network connectivity, have not been studied carefully enough yet. To address the three problems above, the authors derive a better measure of network connectivity for large networks and a new strategy that can increase/decrease network connectivity the most, and propose a spectral density algorithmoptimizing the connectivity of large networks, which is able to indicate the impact on the structural properties of a network while increasing/decreasing its connectivity, providing us a guided optimization of network connectivity. In other words, our algorithm can optimize not only the connectivity of a large network but also its structural features. Meanwhile, our new findings about spectral density are also concluded in this paper. In addition, we may also apply this algorithm to solve all eigenvalues of an N × N matrix, with a low complexity of O(N2) at most.

A brief introduction of characteristic statistic algorithm (CSA) is given in the paper, which is a new global optimizationalgorithm to solve the problem of PWR in-core fuel management optimization. CSA is modified by the adoption of back propagation neural network and fast local adjustment. Then the modified CSA is applied to PWR Equilibrium Cycle Reloading Optimization, and the corresponding optimization code of CSA-DYW is developed. CSA-DYW is used to optimize the equilibrium cycle of 18 month reloading of Daya bay nuclear plant Unit 1 reactor. The results show that CSA-DYW has high efficiency and good global performance on PWR Equilibrium Cycle Reloading Optimization. (authors)

With the development of wireless sensor networks, certain network problems have become more prominent, such as limited node resources, low data transmission security, and short network life cycles. To solve these problems effectively, it is important to design an efficient and trusted secure routing algorithm for wireless sensor networks. Traditional ant-colony optimizationalgorithms exhibit only local convergence, without considering the residual energy of the nodes and many other problems....

Optimization of OPC recipes is not trivial due to multiple parameters that need tuning and their correlation. Usually, no standard methodologies exist for choosing the initial recipe settings, and in the keyword development phase, parameters are chosen either based on previous learning, vendor recommendations, or to resolve specific problems on particular special constructs. Such approaches fail to holistically quantify the effects of parameters on other or possible new designs, and to an extent are based on the keyword developer's intuition. In addition, when a quick fix is needed for a new design, numerous customization statements are added to the recipe, which make it more complex. The present work demonstrates the application of Genetic Algorithm (GA) technique for optimizing OPC recipes. GA is a search technique that mimics Darwinian natural selection and has applications in various science and engineering disciplines. In this case, GA search heuristic is applied to two problems: (a) an overall OPC recipe optimization with respect to selected parameters and, (b) application of GA to improve printing and via coverage at line end geometries. As will be demonstrated, the optimized recipe significantly reduced the number of ORC violations for case (a). For case (b) line end for various features showed significant printing and filling improvement.

Long-duration surface missions to the Moon and Mars will require bases to accommodate habitats for the astronauts. Transporting the materials and equipment required to build the necessary habitats is costly and difficult. The materials chosen for the habitat walls play a direct role in protection against each of the mentioned hazards. Choosing the best materials, their configuration, and the amount required is extremely difficult due to the immense size of the design region. Clearly, an optimization method is warranted for habitat wall design. Standard optimization techniques are not suitable for problems with such large search spaces; therefore, a habitat wall design tool utilizing genetic algorithms (GAs) has been developed. GAs use a "survival of the fittest" philosophy where the most fit individuals are more likely to survive and reproduce. This habitat design optimization tool is a multiobjective formulation of up-mass, heat loss, structural analysis, meteoroid impact protection, and radiation protection. This Technical Publication presents the research and development of this tool as well as a technique for finding the optimal GA search parameters.

Full Text Available Optimization of machining processes not only increases machining efficiency and economics, but also the end product quality. In recent years, among the traditional optimization methods, stochastic direct search optimization methods such as meta-heuristic algorithms are being increasingly applied for solving machining optimization problems. Their ability to deal with complex, multi-dimensional and ill-behaved optimization problems made them the preferred optimization tool by most researchers and practitioners. This paper introduces the use of pattern search (PS algorithm, as a deterministic direct search optimization method, for solving machining optimization problems. To analyze the applicability and performance of the PS algorithm, six case studies of machining optimization problems, both single and multi-objective, were considered. The PS algorithm was employed to determine optimal combinations of machining parameters for different machining processes such as abrasive waterjet machining, turning, turn-milling, drilling, electrical discharge machining and wire electrical discharge machining. In each case study the optimization solutions obtained by the PS algorithm were compared with the optimization solutions that had been determined by past researchers using meta-heuristic algorithms. Analysis of obtained optimization results indicates that the PS algorithm is very applicable for solving machining optimization problems showing good competitive potential against stochastic direct search methods such as meta-heuristic algorithms. Specific features and merits of the PS algorithm were also discussed.

This parameter is taken as the threshold of neuron for learning of neural network. This algorithm is tested with three benchmark datasets and ... Author Affiliations. OM PRAKASH PATEL1 ARUNA TIWARI. Department of Computer Science and Engineering, Indian Institute of Technology Indore, Indore 453552, India ...

Full Text Available An age of network has been living for the last decades. The information technologies have been used by hundreds of millions of users. These technologies are enabling to connect businesses and economic activities. One of the characteristics of the networked economy is the amount of data that produced due to the interlinking of firms, individuals, processes by businesses, and economic activities. Another issue with the networked economy is the complexity of the data. Extraction of the knowledge from the networked economy has challenges by the traditional approach since data is large scale, second decentralized, and third they connect many heterogeneous agents. The challenges can be overcome by the new optimization methods including human element or the social interactions with technological infrastructure.

This article describes the design of highly complex physical instruments by using a canonical genetic algorithm (GA). The procedure can be applied to all instrument designs where performance goals can be quantified. It is particularly suited to the optimization of instrument design where local optima in the performance figure of merit are prevalent. Here, a GA is used to evolve the design of the neutron spin-echo spectrometer WASP which is presently being constructed at the Institut Laue-Langevin, Grenoble, France. A comparison is made between this artificial intelligence approach and the traditional manual design methods. We demonstrate that the search of parameter space is more efficient when applying the genetic algorithm, and the GA produces a significantly better instrument design. Furthermore, it is found that the GA increases flexibility, by facilitating the reoptimization of the design after changes in boundary conditions during the design phase. The GA also allows the exploration of 'nonstandard' magnet coil geometries. We conclude that this technique constitutes a powerful complementary tool for the design and optimization of complex scientific apparatus, without replacing the careful thought processes employed in traditional design methods

This article describes the design of highly complex physical instruments by using a canonical genetic algorithm (GA). The procedure can be applied to all instrument designs where performance goals can be quantified. It is particularly suited to the optimization of instrument design where local optima in the performance figure of merit are prevalent. Here, a GA is used to evolve the design of the neutron spin-echo spectrometer WASP which is presently being constructed at the Institut Laue-Langevin, Grenoble, France. A comparison is made between this artificial intelligence approach and the traditional manual design methods. We demonstrate that the search of parameter space is more efficient when applying the genetic algorithm, and the GA produces a significantly better instrument design. Furthermore, it is found that the GA increases flexibility, by facilitating the reoptimization of the design after changes in boundary conditions during the design phase. The GA also allows the exploration of "nonstandard" magnet coil geometries. We conclude that this technique constitutes a powerful complementary tool for the design and optimization of complex scientific apparatus, without replacing the careful thought processes employed in traditional design methods.

Full Text Available In wireless sensor networks (WSNs, there generally exist many different objective functions to be optimized. In this paper, we propose a stochastic multiobjective optimization approach to solve such kind of problem. We first formulate a general multiobjective optimization problem. We then decompose the optimization formulation through Lagrange dual decomposition and adopt the stochastic quasigradient algorithm to solve the primal-dual problem in a distributed way. We show theoretically that our algorithm converges to the optimal solution of the primal problem by using the knowledge of stochastic programming. Furthermore, the formulation provides a general stochastic multiobjective optimization framework for WSNs. We illustrate how the general framework works by considering an example of the optimal rate allocation problem in multipath WSNs with time-varying channel. Extensive simulation results are given to demonstrate the effectiveness of our algorithm.

The book presents recently developed efficient metaheuristic optimizationalgorithms and their applications for solving various optimization problems in civil engineering. The concepts can also be used for optimizing problems in mechanical and electrical engineering.

Full Text Available Nonlinear model predictive control has been established as a powerful methodology to provide feedback for dynamic processes over the last decades. In practice it is usually combined with parameter and state estimation techniques, which allows to cope with uncertainty on many levels. To reduce the uncertainty it has also been suggested to include optimal experimental design into the sequential process of estimation and control calculation. Most of the focus so far was on dual control approaches, i.e., on using the controls to simultaneously excite the system dynamics (learning as well as minimizing a given objective (performing. We propose a new algorithm, which sequentially solves robust optimal control, optimal experimental design, state and parameter estimation problems. Thus, we decouple the control and the experimental design problems. This has the advantages that we can analyze the impact of measurement timing (sampling independently, and is practically relevant for applications with either an ethical limitation on system excitation (e.g., chemotherapy treatment or the need for fast feedback. The algorithm shows promising results with a 36% reduction of parameter uncertainties for the Lotka-Volterra fishing benchmark example.

This book provides a single-source reference to routing algorithms for Networks-on-Chip (NoCs), as well as in-depth discussions of advanced solutions applied to current and next generation, many core NoC-based Systems-on-Chip (SoCs). After a basic introduction to the NoC design paradigm and architectures, routing algorithms for NoC architectures are presented and discussed at all abstraction levels, from the algorithmic level to actual implementation. Coverage emphasizes the role played by the routing algorithm and is organized around key problems affecting current and next generation, many-core SoCs. A selection of routing algorithms is included, specifically designed to address key issues faced by designers in the ultra-deep sub-micron (UDSM) era, including performance improvement, power, energy, and thermal issues, fault tolerance and reliability. · Provides a comprehensive overview of routing algorithms for Networks-on-Chip and NoC-based, manycore systems; · Describe...

In order to meet demanding performance objectives in Long Term Evolution (LTE) networks, it is mandatory to implement highly efficient, autonomic self-optimization and configuration processes. Self-optimization processes have already been studied in second generation (2G) and third generation (3G) networks, typically with the objective of improving radio coverage and channel capacity. The 3rd Generation Partnership Project (3GPP) standard for LTE self-organization of networks (SON) provides guidelines on self-configuration of physical cell ID and neighbor relation function and self-optimization for mobility robustness, load balancing, and inter-cell interference reduction. While these are very important from an optimization perspective of local phenomenon (i.e., the eNodeB's interaction with its neighbors), it is also essential to architect control algorithms to optimize the network as a whole. In this paper, we propose a Celnet Xplorer-based SON architecture that allows detailed analysis of network performan...

Four measurement algorithms for flow enforcement in asynchronous transfer mode (ATM) networks are presented. The algorithms are the leaky bucket, the rectangular sliding window, the triangular sliding window, and the exponentially weighted moving average. A comparison, based partly on teletraffic...... theory and partly on signal processing theory, is carried out. It is seen that the time constant involved increases with the increasing burstiness of the connection. It is suggested that the RMS measurement bandwidth be used to dimension linear algorithms for equal flow enforcement characteristics....... Implementations are proposed on the block diagram level, and dimensioning examples are carried out when flow enforcing a renewal-type connection using the four algorithms. The corresponding hardware demands are estimated aid compared...

In this work a heuristic optimizationalgorithm known as the Fruit fly OptimizationAlgorithm is applied to antenna design problems. The original formulation of the algorithm is presented and it is adapted to array factor and horn antenna optimization problems. Specifically, it is applied to the array factor synthesis of uniformly-fed, non-equispaced arrays and to the profile optimization of multimode horn antennas. Several numerical examples are presented and the obtained results are compare...

In this work a heuristic optimizationalgorithm known as the Fruit fly OptimizationAlgorithm is applied to antenna design problems. The original formulation of the algorithm is presented and it is adapted to array factor and horn antenna optimization problems. Specifically, it is applied to the array factor synthesis of uniformly-fed, non-equispaced arrays and to the profile optimization of multimode horn antennas. Several numerical examples are presented and the obtained results are compare...

A swarming mobile sensor network is comprised of a swarm of wirelessly connected mobile robots equipped with various sensors. Such a network can be applied in an uncertain environment for services such as cooperative navigation and exploration, object identification and information gathering. One of the most advantageous properties of the swarming wireless sensor network is that mobile nodes can work cooperatively to organize an ad-hoc network and optimize the network link capacity to maximize the transmission of gathered data from a source to a target. This paper describes a new method of link optimization of swarming mobile sensor networks. The new method is based on combination of the artificial potential force guaranteeing connectivities of the mobile sensor nodes and the max-flow min-cut theorem of graph theory ensuring optimization of the network link capacity. The developed algorithm is demonstrated and evaluated in simulation.

Full Text Available A swarming mobile sensor network is comprised of a swarm of wirelessly connected mobile robots equipped with various sensors. Such a network can be applied in an uncertain environment for services such as cooperative navigation and exploration, object identification and information gathering. One of the most advantageous properties of the swarming wireless sensor network is that mobile nodes can work cooperatively to organize an ad-hoc network and optimize the network link capacity to maximize the transmission of gathered data from a source to a target. This paper describes a new method of link optimization of swarming mobile sensor networks. The new method is based on combination of the artificial potential force guaranteeing connectivities of the mobile sensor nodes and the max-flow min-cut theorem of graph theory ensuring optimization of the network link capacity. The developed algorithm is demonstrated and evaluated in simulation.

Ant colony optimization (ACO) algorithms often fall into the local optimal solution and have lower search efficiency for solving the travelling salesman problem (TSP). According to these shortcomings, this paper proposes a universal optimization strategy for updating the pheromone matrix in the ACO algorithms. The new optimization strategy takes advantages of the unique feature of critical paths reserved in the process of evolving adaptive networks of the Physarum-inspired mathematical model (PMM). The optimizedalgorithms, denoted as PMACO algorithms, can enhance the amount of pheromone in the critical paths and promote the exploitation of the optimal solution. Experimental results in synthetic and real networks show that the PMACO algorithms are more efficient and robust than the traditional ACO algorithms, which are adaptable to solve the TSP with single or multiple objectives. Meanwhile, we further analyse the influence of parameters on the performance of the PMACO algorithms. Based on these analyses, the best values of these parameters are worked out for the TSP. (paper)

In this paper, a Hybrid Genetic Based Optimization for Multicast Routing algorithm is proposed. The proposed algorithm uses the best features of Genetic Algorithm (GA) and particle swarm optimization (PSO) to improve the solution. Simulations were conducted by varying number of mobile nodes and results compared with ...

Marriage in honey bees optimization (MBO) is a metaheuristic optimizationalgorithm developed by inspiration of the mating and fertilization process of honey bees and is a kind of swarm intelligence optimizations. In this study we propose improved marriage in honey bees optimization (IMBO) by adding Levy flight algorithm for queen mating flight and neighboring for worker drone improving. The IMBO algorithm's performance and its success are tested on the well-known six unconstrained test functions and compared with other metaheuristic optimizationalgorithms.

This paper proposes an optimization platform based on Genetic Algorithm, where the main components of the electrical system of a wind farm and key technical specifications are used as input parameters and the topology of the electrical system is to be optimized for a minimum cost and high...... reliability. A method to encode and decode an electrical system is studied. The reliability evaluation for a given network is also investigated. Genetic Algorithm is implemented to find the optimum network design for a large DC wind farm. It is concluded that different topologies may cause very different cost...... and reliability, and the Genetic Algorithm is capable of finding the optimum solution....

Full Text Available In this work a heuristic optimizationalgorithm known as the Fruit fly OptimizationAlgorithm is applied to antenna design problems. The original formulation of the algorithm is presented and it is adapted to array factor and horn antenna optimization problems. Specifically, it is applied to the array factor synthesis of uniformly-fed, non-equispaced arrays and to the profile optimization of multimode horn antennas. Several numerical examples are presented and the obtained results are compared with those provided by a deterministic optimization based on a simplex method and another well-known heuristic approach, the Genetic Algorithm.

Structural balance is a large area of study in signed networks, and it is intrinsically a global property of the whole network. Computing global structural balance in signed networks, which has attracted some attention in recent years, is to measure how unbalanced a signed network is and it is a nondeterministic polynomial-time hard problem. Many approaches are developed to compute global balance. However, the results obtained by them are partial and unsatisfactory. In this study, the computation of global structural balance is solved as an optimization problem by using the Memetic Algorithm. The optimizationalgorithm, named Meme-SB, is proposed to optimize an evaluation function, energy function, which is used to compute a distance to exact balance. Our proposed algorithm combines Genetic Algorithm and a greedy strategy as the local search procedure. Experiments on social and biological networks show the excellent effectiveness and efficiency of the proposed method.

Full Text Available Printed circuit antennas have been used for the detection of electromagnetic radiation at a wide range of frequencies that go from radio frequencies (RF up to optical frequencies. The design of printed antennas at optical frequencies has been done by using design rules derived from the radio frequency domain which do not take into account the dispersion of material parameters at optical frequencies. This can make traditional RF antenna design not suitable for optical antenna design. This work presents the results of using a genetic algorithm (GA for obtaining an optimized geometry (unconventional geometries that may be used as optical regime antennas to capture electromagnetic waves. The radiation patterns and optical properties of the GA generated geometries were compared with the conventional dipole geometry. The characterizations were conducted via finite element method (FEM computational simulations.

Genetic algorithms, which simulate evolution in natural systems, have been used to find solutions to optimization problems that seem intractable to standard approaches. In this study, the feasibility of using a GA to find an optimum, fixed profile nozzle for a pulse detonation engine (PDE) is demonstrated. The objective was to maximize impulse during the detonation wave passage and blow-down phases of operation. Impulse of each profile variant was obtained by using the CFD code Mozart/2.0 to simulate the transient flow. After 7 generations, the method has identified a nozzle profile that certainly is a candidate for optimum solution. The constraints on the generality of this possible solution remain to be clarified.

High-energy physics experiments rely on reconstruction of the trajectories of particles produced at the interaction point. This is a challenging task, especially in the high track multiplicity environment generated by p-p collisions at the LHC energies. A typical event includes hundreds of signal examples (interesting decays) and a significant amount of noise (uninteresting examples). This work describes a modification of the Artificial Retina algorithm for fast track finding: numerical optimization methods were adopted for fast local track search. This approach allows for considerable reduction of the total computational time per event. Test results on simplified simulated model of LHCb VELO (VErtex LOcator) detector are presented. Also this approach is well-suited for implementation of paralleled computations as GPGPU which look very attractive in the context of upcoming detector upgrades.

The learning or training process of neural networks entails the task of finding the most optimal set of parameters, which includes translation vectors, dilation parameter, synaptic weights, and bias terms. Apart from the traditional gradient descent-based methods, metaheuristic methods can also be used for this learning purpose. Since the inception of genetic algorithm half a century ago, the last decade witnessed the explosion of a variety of novel metaheuristic algorithms, such as harmony search algorithm, bat algorithm, and whale optimizationalgorithm. Despite the proof of the no free lunch theorem in the discipline of optimization, a survey in the literature of machine learning gives contrasting results. Some researchers report that certain metaheuristic algorithms are superior to the others, whereas some others argue that different metaheuristic algorithms give comparable performance. As such, this paper aims to investigate if a certain metaheuristic algorithm will outperform the other algorithms. In this work, three metaheuristic algorithms, namely genetic algorithms, particle swarm optimization, and harmony search algorithm are considered. The algorithms are incorporated in the learning of neural networks and their classification results on the benchmark UCI machine learning data sets are compared. It is found that all three metaheuristic algorithms give similar and comparable performance, as captured in the average overall classification accuracy. The results corroborate the findings reported in the works done by previous researchers. Several recommendations are given, which include the need of statistical analysis to verify the results and further theoretical works to support the obtained empirical results.

Currently, overlay measurements are characterized by "recipe", which defines both physical parameters such as focus, illumination et cetera, and also the software parameters such as algorithm to be used and regions of interest. Setting up these recipes requires both engineering time and wafer availability on an overlay tool, so reducing these requirements will result in higher tool productivity. One of the significant challenges to automating this process is that the parameters are highly and complexly correlated. At the same time, a high level of traceability and transparency is required in the recipe creation process, so a technique that maintains its decisions in terms of well defined physical parameters is desirable. Running time should be short, given the system (automatic recipe creation) is being implemented to reduce overheads. Finally, a failure of the system to determine acceptable parameters should be obvious, so a certainty metric is also desirable. The complex, nonlinear interactions make solution by an expert system difficult at best, especially in the verification of the resulting decision network. The transparency requirements tend to preclude classical neural networks and similar techniques. Genetic algorithms and other "global minimization" techniques require too much computational power (given system footprint and cost requirements). A Bayesian network, however, provides a solution to these requirements. Such a network, with appropriate priors, can be used during recipe creation / optimization not just to select a good set of parameters, but also to guide the direction of search, by evaluating the network state while only incomplete information is available. As a Bayesian network maintains an estimate of the probability distribution of nodal values, a maximum-entropy approach can be utilized to obtain a working recipe in a minimum or near-minimum number of steps. In this paper we discuss the potential use of a Bayesian network in such a capacity

Full Text Available The quest for an efficient nature-inspired optimization technique has continued over the last few decades. In this paper, a hybrid nature-inspired optimization technique has been proposed. The hybrid algorithm has been constructed using Mean Grey Wolf Optimizer (MGWO and Whale OptimizerAlgorithm (WOA. We have utilized the spiral equation of Whale OptimizerAlgorithm for two procedures in the Hybrid Approach GWO (HAGWO algorithm: (i firstly, we used the spiral equation in Grey Wolf OptimizerAlgorithm for balance between the exploitation and the exploration process in the new hybrid approach; and (ii secondly, we also applied this equation in the whole population in order to refrain from the premature convergence and trapping in local minima. The feasibility and effectiveness of the hybrid algorithm have been tested by solving some standard benchmarks, XOR, Baloon, Iris, Breast Cancer, Welded Beam Design, Pressure Vessel Design problems and comparing the results with those obtained through other metaheuristics. The solutions prove that the newly existing hybrid variant has higher stronger stability, faster convergence rate and computational accuracy than other nature-inspired metaheuristics on the maximum number of problems and can successfully resolve the function of constrained nonlinear optimization in reality.

We present a novel, Hybrid Aggressive Space Mapping (HASM) optimizationalgorithm. HASM is a hybrid approach exploiting both the Trust Region Aggressive Space Mapping (TRASM) algorithm and direct optimization. It does not assume that the final space-mapped design is the true optimal design...

Niching methods extend genetic algorithms and permit the investigation of multiple optimal solutions in the search space. In this paper, we review and discuss various strategies of niching for optimization in electromagnetics. Traditional mathematical problems and an electromagnetic benchmark are solved using niching genetic algorithms to show their interest in real world optimization.

Full Text Available There is a growing interest in finding a global optimal path in transportation networks particularly when the network suffers from unexpected disturbance. This paper studies the problem of finding a global optimal path to guarantee a given probability of arriving on time in a network with uncertainty, in which the travel time is stochastic instead of deterministic. Traditional path finding methods based on least expected travel time cannot capture the network user’s risk-taking behaviors in path finding. To overcome such limitation, the reliable path finding algorithms have been proposed but the convergence of global optimum is seldom addressed in the literature. This paper integrates the K-shortest path algorithm into Backtracking method to propose a new path finding algorithm under uncertainty. The global optimum of the proposed method can be guaranteed. Numerical examples are conducted to demonstrate the correctness and efficiency of the proposed algorithm.

Full Text Available The aim of this work is to investigate new approaches using methods based on statistics and geo-statistics for spatio-temporal optimization of groundwater monitoring networks. The formulated and integrated methods were tested with the groundwater quality data set of Bitterfeld/Wolfen, Germany. Spatially, the monitoring network was optimized using geo-statistical methods. Temporal optimization of the monitoring network was carried out using Sen’s method (1968. For geostatistical networkoptimization, a geostatistical spatio-temporal algorithm was used to identify redundant wells in 2- and 2.5-D Quaternary and Tertiary aquifers. Influences of interpolation block width, dimension, contaminant association, groundwater flow direction and aquifer homogeneity on statistical and geostatistical methods for monitoring networkoptimization were analysed. The integrated approach shows 37% and 28% redundancies in the monitoring network in Quaternary aquifer and Tertiary aquifer respectively. The geostatistical method also recommends 41 and 22 new monitoring wells in the Quaternary and Tertiary aquifers respectively. In temporal optimization, an overall optimized sampling interval was recommended in terms of lower quartile (238 days, median quartile (317 days and upper quartile (401 days in the research area of Bitterfeld/Wolfen. Demonstrated methods for improving groundwater monitoring network can be used in real monitoring networkoptimization with due consideration given to influencing factors.

To investigate the performance of quantum information tasks on networks whose topology changes in time, we study the spatial search algorithm by continuous time quantum walk to find a marked node on a random temporal network. We consider a network of n nodes constituted by a time-ordered sequence of Erdös-Rényi random graphs G (n ,p ), where p is the probability that any two given nodes are connected: After every time interval τ , a new graph G (n ,p ) replaces the previous one. We prove analytically that, for any given p , there is always a range of values of τ for which the running time of the algorithm is optimal, i.e., O (√{n }), even when search on the individual static graphs constituting the temporal network is suboptimal. On the other hand, there are regimes of τ where the algorithm is suboptimal even when each of the underlying static graphs are sufficiently connected to perform optimal search on them. From this first study of quantum spatial search on a time-dependent network, it emerges that the nontrivial interplay between temporality and connectivity is key to the algorithmic performance. Moreover, our work can be extended to establish high-fidelity qubit transfer between any two nodes of the network. Overall, our findings show that one can exploit temporality to achieve optimal quantum information tasks on dynamical random networks.

To investigate the performance of quantum information tasks on networks whose topology changes in time, we study the spatial search algorithm by continuous time quantum walk to find a marked node on a random temporal network. We consider a network of n nodes constituted by a time-ordered sequence of Erdös-Rényi random graphs G(n,p), where p is the probability that any two given nodes are connected: After every time interval τ, a new graph G(n,p) replaces the previous one. We prove analytically that, for any given p, there is always a range of values of τ for which the running time of the algorithm is optimal, i.e., O(sqrt[n]), even when search on the individual static graphs constituting the temporal network is suboptimal. On the other hand, there are regimes of τ where the algorithm is suboptimal even when each of the underlying static graphs are sufficiently connected to perform optimal search on them. From this first study of quantum spatial search on a time-dependent network, it emerges that the nontrivial interplay between temporality and connectivity is key to the algorithmic performance. Moreover, our work can be extended to establish high-fidelity qubit transfer between any two nodes of the network. Overall, our findings show that one can exploit temporality to achieve optimal quantum information tasks on dynamical random networks.

Full Text Available Audio segmentation is a basis for multimedia content analysis which is the most important and widely used application nowadays. An optimized audio classification and segmentation algorithm is presented in this paper that segments a superimposed audio stream on the basis of its content into four main audio types: pure-speech, music, environment sound, and silence. An algorithm is proposed that preserves important audio content and reduces the misclassification rate without using large amount of training data, which handles noise and is suitable for use for real-time applications. Noise in an audio stream is segmented out as environment sound. A hybrid classification approach is used, bagged support vector machines (SVMs with artificial neural networks (ANNs. Audio stream is classified, firstly, into speech and nonspeech segment by using bagged support vector machines; nonspeech segment is further classified into music and environment sound by using artificial neural networks and lastly, speech segment is classified into silence and pure-speech segments on the basis of rule-based classifier. Minimum data is used for training classifier; ensemble methods are used for minimizing misclassification rate and approximately 98% accurate segments are obtained. A fast and efficient algorithm is designed that can be used with real-time multimedia applications.

ABSTRACT In the fields which require finding the most appropriate value, optimization became a vital approach to employ effective solutions. With the use of optimization techniques, many different fields in the modern life have found solutions to their real-world based problems. In this context, classical optimization techniques have had an important popularity. But after a while, more advanced optimization problems required the use of more effective techniques. At this point, Computer Sc...

In the fields which require finding the most appropriate value, optimization became a vital approach to employ effective solutions. With the use of optimization techniques, many different fields in the modern life have found solutions to their real-world based problems. In this context, classical optimization techniques have had an important popularity. But after a while, more advanced optimization problems required the use of more effective techniques. At this point, Computer Science took an...

Full Text Available Abstract We propose ZAP, an algorithm for the distributed channel assignment in cognitive radio (CR networks. CRs are capable of identifying underutilized licensed bands of the spectrum, allowing their reuse by secondary users without interfering with primary users. In this context, efficient channel assignment is challenging as ideally it must be simple, incur acceptable communication overhead, provide timely response, and be adaptive to accommodate frequent changes in the network. Another challenge is the optimization of network capacity through interference minimization. In contrast to related work, ZAP addresses these challenges with a fully distributed approach based only on local (neighborhood knowledge, while significantly reducing computational costs and the number of messages required for channel assignment. Simulations confirm the efficiency of ZAP in terms of (i the performance tradeoff between different metrics and (ii the fast achievement of a suitable assignment solution regardless of network size and density.

Full Text Available With the deepening of the research on complex networks, the method of detecting and classifying social network is springing up. In this essay, the basic particle swarm algorithm is improved based on the GN algorithm. Modularity is taken as a measure of community division [1]. In view of the dynamic network community division, scrolling calculation method is put forward. Experiments show that using the improved particle swarm optimizationalgorithm can improve the accuracy of the community division and can also get higher value of the modularity in the dynamic community

Recommender systems are a vital tool that helps us to overcome the information overload problem. They are being used by most e-commerce web sites and attract the interest of a broad scientific community. A recommender system uses data on users' past preferences to choose new items that might be appreciated by a given individual user. While many approaches to recommendation exist, the approach based on a network representation of the input data has gained considerable attention in the past. We review here a broad range of network-based recommendation algorithms and for the first time compare their performance on three distinct real datasets. We present recommendation topics that go beyond the mere question of which algorithm to use-such as the possible influence of recommendation on the evolution of systems that use it-and finally discuss open research directions and challenges.

Full Text Available In this paper, a Cellular Automata method is applied for the optimal design of sewer networks. The solution of sewer networkoptimization problems requires the determination of pipe diameters and average pipe cover depths, minimizing the total cost of the sewer network subject to operational constraints. In this paper, the network nodes and upstream and downstream pipe cover depths are considered as CA cells and cell states, respectively, and the links around each cell are taken into account as neighborhood. The proposed method is a general and flexible method for the optimization of sewer networks as it can be used to optimally design both gravity and pumped network due to the use of pipe nodal cover depths as the decision variables. The proposed method is tested against two gravitational sewer networks and the comparison of results with other methods such as Genetic algorithm, Cellular Automata, Ant Colony OptimizationAlgorithm and Particle Swarm Optimization show the efficiency and effectiveness of the proposed method.

By employing a recently introduced optimizationalgorithm we construct optimally synchronizable (unweighted) networks for any given scale-free degree distribution. We explore how the optimization process affects degree-degree correlations and observe a generic tendency toward disassortativity. Still, we show that there is not a one-to-one correspondence between synchronizability and disassortativity. On the other hand, we study the nature of optimally un-synchronizable networks, that is, networks whose topology minimizes the range of stability of the synchronous state. The resulting 'pessimal networks' turn out to have a highly assortative string-like structure. We also derive a rigorous lower bound for the Laplacian eigenvalue ratio controlling synchronizability, which helps understanding the impact of degree correlations on network synchronizability

By employing a recently introduced optimizationalgorithm we construct optimally synchronizable (unweighted) networks for any given scale-free degree distribution. We explore how the optimization process affects degree-degree correlations and observe a generic tendency toward disassortativity. Still, we show that there is not a one-to-one correspondence between synchronizability and disassortativity. On the other hand, we study the nature of optimally un-synchronizable networks, that is, networks whose topology minimizes the range of stability of the synchronous state. The resulting 'pessimal networks' turn out to have a highly assortative string-like structure. We also derive a rigorous lower bound for the Laplacian eigenvalue ratio controlling synchronizability, which helps understanding the impact of degree correlations on network synchronizability.

The robustness research of networked systems has drawn widespread attention in the past decade, and one of the central topics is to protect the network from external attacks through allocating appropriate defense resource to different nodes. In this paper, we apply a specific particle swarm optimization (PSO) algorithm to optimize the defense resource allocation in scale-free networks. Results reveal that PSO based resource allocation shows a higher robustness than other resource allocation strategies such as uniform, degree-proportional, and betweenness-proportional allocation strategies. Furthermore, we find that assigning less resource to middle-degree nodes under small-scale attack while more resource to low-degree nodes under large-scale attack is conductive to improving the network robustness. Our work provides an insight into the optimal defense resource allocation pattern in scale-free networks and is helpful for designing a more robust network.

Full Text Available In order to enhance the convergence capability of the central force optimization (CFO algorithm, an adaptive central force optimization (ACFO algorithm is presented by introducing an adaptive weight and defining an adaptive gravitational constant. The adaptive weight and gravitational constant are selected based on the stability theory of discrete time-varying dynamic systems. The convergence capability of ACFO algorithm is compared with the other improved CFO algorithm and evolutionary-based algorithm using 23 unimodal and multimodal benchmark functions. Experiments results show that ACFO substantially enhances the performance of CFO in terms of global optimality and solution accuracy.

The purpose of this paper is to improve theVelocity Update Relaxation Particle Swarm Optimizationalgorithm (VURPSO). The improved algorithm is called Adaptive VURPSO (AVURPSO) algorithm. Then, an optimal design of a Proportional-Integral-Derivative (PID) controller is obtained using the AVURPSO algorithm. An adaptive momentum factor is used to regulate a trade-off between the global and the local exploration abilities in the proposed algorithm. This operation helps the system to reach the optimal solution quickly and saves the computation time. Comparisons on the optimal PID controller design confirm the superiority of AVURPSO algorithm to the optimizationalgorithms mentioned in this paper namely the VURPSO algorithm, the Ant Colony algorithm, and the conventional approach. Comparisons on the speed of convergence confirm that the proposed algorithm has a faster convergence in a less computation time to yield a global optimum value. The proposed AVURPSO can be used in the diverse areas of optimization problems such as industrial planning, resource allocation, scheduling, decision making, pattern recognition and machine learning. The proposed AVURPSO algorithm is efficiently used to design an optimal PID controller.

Stochastic methods have gained some popularity in global optimization in that most of them do not assume the cost functions to be differentiable. They have capabilities to avoid being trapped by local optima, and may converge even faster than gradient-based optimization methods on some problems. The present paper proposes an optimization method, which reduces the search space by means of densification curves, coupled with the dynamic canonical descent algorithm. The performances of the new method are shown on several known problems classically used for testing optimizationalgorithms, and proved to outperform competitive algorithms such as simulated annealing and genetic algorithms.

National Aeronautics and Space Administration — NASA is developing algorithms and methodologies for efficient air-traffic management. Several researchers have adopted an optimization framework for solving problems...

The natural gas networks require very important investments to cope with a still growing demand and to satisfy the new regulatory constraints. The gas market deregulation imposed to the gas network operators, first, transparency rules of a natural monopoly to justify their costs and ultimately their tariffs, and, second, market fluidity objectives in order to facilitate access for competition to the end-users. These major investments are the main reasons for the use of optimization techniques aiming at reducing the costs. Due to the discrete choices (investment location, limited choice of additional capacities, timing) crossed with physical non linear constraints (flow/pressures relations in the pipe or operating ranges of compressors), the programs to solve are Large Mixed Non Linear Programs (MINLP). As these types of programs are known to be hard to solve exactly in polynomial times (NP-hard), advanced optimization methods have to be implemented to obtain realistic results. The objectives of this thesis are threefold. First, one states several investment problems modeling of natural gas networks from industrial world motivations. Second, one identifies the most suitable methods and algorithms to the formulated problems. Third, one exposes the main advantages and drawbacks of these methods with the help of numerical applications on real cases. (author)

The availability of whole-cell level metabolic networks of high quality has made it possible to develop a predictive understanding of bacterial metabolism. Using the optimization framework of flux balance analysis, I investigate metabolic response and activity patterns to variations in the availability of nutrient and chemical factors such as oxygen and ammonia by simulating 30,000 random cellular environments. The distribution of reaction fluxes is heavy-tailed for the bacteria H. pylori and E. coli, and the eukaryote S. cerevisiae. While the majority of flux balance investigations have relied on implementations of the simplex method, it is necessary to use interior-point optimizationalgorithms to adequately characterize the full range of activity patterns on metabolic networks. The interior-point activity pattern is bimodal for E. coli and S. cerevisiae, suggesting that most metabolic reaction are either in frequent use or are rarely active. The trimodal activity pattern of H. pylori indicates that a group of its metabolic reactions (20%) are active in approximately half of the simulated environments. Constructing the high-flux backbone of the network for every environment, there is a clear trend that the more frequently a reaction is active, the more likely it is a part of the backbone. Finally, I briefly discuss the predicted activity patterns of the central-carbon metabolic pathways for the sample of random environments.

Full Text Available Multicast routing is the basic demand to provide QOS (Quality of service in multimedia streaming on peer to peer networks. Making multicast trees optimizing their delay cost and considering nodal and links limited bandwidth (load balance constraints is a NP-hard (Nondeterministic Polynomial time hard problem. In this paper we have used Co-evolutionary Algorithm to make multicast trees with optimized average delay from source to the clients considering the limited capacity of nodes and links in application layer. The numeric results obtained are shown that the costs have been much improved comparing with other existent non-GA (Genetic Algorithm approaches. Also we have used only a portion of every nodal outage degree and this has improved the results comparing to use of the entire outage degree.

We formulate a technique for the detection of functional clusters in discrete event data. The advantage of this algorithm is that no prior knowledge of the number of functional groups is needed, as our procedure progressively combines data traces and derives the optimal clustering cutoff in a simple and intuitive manner through the use of surrogate data sets. In order to demonstrate the power of this algorithm to detect changes in network dynamics and connectivity, we apply it to both simulated neural spike train data and real neural data obtained from the mouse hippocampus during exploration and slow-wave sleep. Using the simulated data, we show that our algorithm performs better than existing methods. In the experimental data, we observe state-dependent clustering patterns consistent with known neurophysiological processes involved in memory consolidation.

This paper shows findings of the application of Particle Swarm Optimization (PSO) algorithm in optimizing an Artificial Neural Network that could categorize between ripeness and unripeness stage of citrus suhuensis. The algorithm would adjust the network connections weights and adapt its values during training for best results at the output. Initially, citrus suhuensis fruit’s skin is measured using optically non-destructive method via spectrometer. The spectrometer would transmit VIS (visible spectrum) photonic light radiation to the surface (skin of citrus) of the sample. The reflected light from the sample’s surface would be received and measured by the same spectrometer in terms of reflectance percentage based on VIS range. These measured data are used to train and test the best optimized ANN model. The accuracy is based on receiver operating characteristic (ROC) performance. The result outcomes from this investigation have shown that the achieved accuracy for the optimized is 70.5% with a sensitivity and specificity of 60.1% and 80.0% respectively.

.... A genetic algorithm (GA) was employed to optimize the design of the space launch vehicle. A cost model was incorporated into the optimization process with the goal of minimizing the overall vehicle cost...

Global optimization is challenging to solve due to its nonlinearity and multimodality. Traditional algorithms such as the gradient-based methods often struggle to deal with such problems and one of the current trends is to use metaheuristic algorithms. In this paper, a novel hybrid population-based global optimizationalgorithm, called hybrid firefly algorithm (HFA), is proposed by combining the advantages of both the firefly algorithm (FA) and differential evolution (DE). FA and DE are executed in parallel to promote information sharing among the population and thus enhance searching efficiency. In order to evaluate the performance and efficiency of the proposed algorithm, a diverse set of selected benchmark functions are employed and these functions fall into two groups: unimodal and multimodal. The experimental results show better performance of the proposed algorithm compared to the original version of the firefly algorithm (FA), differential evolution (DE) and particle swarm optimization (PSO) in the sense of avoiding local minima and increasing the convergence rate.

traversal is assigned a score that is distributed among the road segments covered by the route according to a score distribution model. The query returns the road segment(s) with the highest score. To achieve low latency, it is essential to prune the very large search space. We propose two algorithms...... that are shown empirically to be scalable. Given a road network, a set of existing facilities, and a collection of customer route traversals, an optimal segment query returns the optimal road network segment(s) for a new facility. We propose a practical framework for computing this query, where each route...... that adopt different approaches to computing the query. Algorithm AUG uses graph augmentation, and ITE uses iterative road-network partitioning. Empirical studies with real data sets demonstrate that the algorithms are capable of offering high performance in realistic settings....

Full Text Available Evolutionary algorithms (EAs are population-based global search methods. They have been successfully applied to many complex optimization problems. However, EAs are frequently incapable of finding a convergence solution in default of local search mechanisms. Memetic Algorithms (MAs are hybrid EAs that combine genetic operators with local search methods. With global exploration and local exploitation in search space, MAs are capable of obtaining more high-quality solutions. On the other hand, mixed-integer hybrid differential evolution (MIHDE, as an EA-based search algorithm, has been successfully applied to many mixed-integer optimization problems. In this paper, a memetic algorithm based on MIHDE is developed for solving mixed-integer optimization problems. However, most of real-world mixed-integer optimization problems frequently consist of equality and/or inequality constraints. In order to effectively handle constraints, an evolutionary Lagrange method based on memetic algorithm is developed to solve the mixed-integer constrained optimization problems. The proposed algorithm is implemented and tested on two benchmark mixed-integer constrained optimization problems. Experimental results show that the proposed algorithm can find better optimal solutions compared with some other search algorithms. Therefore, it implies that the proposed memetic algorithm is a good approach to mixed-integer optimization problems.

This textbook provides a comprehensive introduction to nature-inspired metaheuristic methods for search and optimization, including the latest trends in evolutionary algorithms and other forms of natural computing. Over 100 different types of these methods are discussed in detail. The authors emphasize non-standard optimization problems and utilize a natural approach to the topic, moving from basic notions to more complex ones. An introductory chapter covers the necessary biological and mathematical backgrounds for understanding the main material. Subsequent chapters then explore almost all of the major metaheuristics for search and optimization created based on natural phenomena, including simulated annealing, recurrent neural networks, genetic algorithms and genetic programming, differential evolution, memetic algorithms, particle swarm optimization, artificial immune systems, ant colony optimization, tabu search and scatter search, bee and bacteria foraging algorithms, harmony search, biomolecular computin...

Full Text Available Optimizationalgorithms are search methods where the goal is to find an optimal solution to a problem, in order to satisfy one or more objective functions, possibly subject to a set of constraints. Studies of social animals and social insects have resulted in a number of computational models of swarm intelligence. Within these swarms their collective behavior is usually very complex. The collective behavior of a swarm of social organisms emerges from the behaviors of the individuals of that swarm. Researchers have developed computational optimization methods based on biology such as Genetic Algorithms, Particle Swarm Optimization, and Ant Colony. The aim of this paper is to describe an optimizationalgorithm called the Bees Algorithm, inspired from the natural foraging behavior of honey bees, to find the optimal solution. The algorithm performs both an exploitative neighborhood search combined with random explorative search. In this paper, after an explanation of the natural foraging behavior of honey bees, the basic Bees Algorithm and its improved versions are described and are implemented in order to optimize several benchmark functions, and the results are compared with those obtained with different optimizationalgorithms. The results show that the Bees Algorithm offering some advantage over other optimization methods according to the nature of the problem.

In Underwater Wireless Sensor Networks (UWSNs), localization is one of most important technologies since it plays a critical role in many applications. Motivated by widespread adoption of localization, in this paper, we present a comprehensive survey of localization algorithms. First, we classify localization algorithms into three categories based on sensor nodes’ mobility: stationary localization algorithms, mobile localization algorithms and hybrid localization algorithms. Moreover, we compare the localization algorithms in detail and analyze future research directions of localization algorithms in UWSNs. PMID:22438752

Full Text Available Nature inspired population based algorithms is a research field which simulates different natural phenomena to solve a wide range of problems. Researchers have proposed several algorithms considering different natural phenomena. Teaching-Learning-based optimization (TLBO is one of the recently proposed population based algorithm which simulates the teaching-learning process of the class room. This algorithm does not require any algorithm-specific control parameters. In this paper, elitism concept is introduced in the TLBO algorithm and its effect on the performance of the algorithm is investigated. The effects of common controlling parameters such as the population size and the number of generations on the performance of the algorithm are also investigated. The proposed algorithm is tested on 35 constrained benchmark functions with different characteristics and the performance of the algorithm is compared with that of other well known optimizationalgorithms. The proposed algorithm can be applied to various optimization problems of the industrial environment.

This paper presents the use a neural network and a micro genetic algorithm to optimize future set-points in existing hydronic floor heating systems for improved energy efficiency. The neural network can be trained to predict the impact of changes in set-points on future room temperatures. Additio...... space is not guaranteed. Evaluation of the performance of multiple neural networks is performed, using different levels of information, and optimization results are presented on a detailed house simulation model....

Full Text Available The Fourth Party Logistics (4PL network faces disruptions of various sorts under the dynamic and complex environment. In order to explore the robustness of the network, the 4PL network design with consideration of random disruptions is studied. The purpose of the research is to construct a 4PL network that can provide satisfactory service to customers at a lower cost when disruptions strike. Based on the definition of β-robustness, a robust optimization model of 4PL network design under disruptions is established. Based on the NP-hard characteristic of the problem, the artificial fish swarm algorithm (AFSA and the genetic algorithm (GA are developed. The effectiveness of the algorithms is tested and compared by simulation examples. By comparing the optimal solutions of the 4PL network for different robustness level, it is indicated that the robust optimization model can evade the market risks effectively and save the cost in the maximum limit when it is applied to 4PL network design.

Full Text Available Training an artificial neural network is an optimization task, since it is desired to find optimal weight sets for a neural network during training process. Traditional training algorithms such as back propagation have some drawbacks such as getting stuck in local minima and slow speed of convergence. This study combines the best features of two algorithms; i.e. Levenberg Marquardt back propagation (LMBP and Cuckoo Search (CS for improving the convergence speed of artificial neural networks (ANN training. The proposed CSLM algorithm is trained on XOR and OR datasets. The experimental results show that the proposed CSLM algorithm has better performance than other similar hybrid variants used in this study.

Full Text Available This paper proposes a new hybrid charged system search (CSS algorithm based state estimation in radial distribution networks in fuzzy framework. The objective of the optimization problem is to minimize the weighted square of the difference between the measured and the estimated quantity. The proposed method of state estimation considers bus voltage magnitude and phase angle as state variable along with some equality and inequality constraints for state estimation in distribution networks. A rule based fuzzy inference system has been designed to control the parameters of the CSS algorithm to achieve better balance between the exploration and exploitation capability of the algorithm. The efficiency of the proposed fuzzy adaptive charged system search (FACSS algorithm has been tested on standard IEEE 33-bus system and Indian 85-bus practical radial distribution system. The obtained results have been compared with the conventional CSS algorithm, weighted least square (WLS algorithm and particle swarm optimization (PSO for feasibility of the algorithm.

The use of neural networks in the reliability optimization field is rare. This paper presents an application of a recent kind of neural networks in a reliability optimization problem for a series system with multiple-choice constraints incorporated at each subsystem, to maximize the system reliability subject to the system budget. The problem is formulated as a nonlinear binary integer programming problem and characterized as an NP-hard problem. Our design of neural network to solve efficiently this problem is based on a quantized Hopfield network. This network allows us to obtain optimal design solutions very frequently and much more quickly than others Hopfield networks

Flower pollination algorithm (FPA) is a new nature-inspired evolutionary algorithm used to solve multi-objective optimization problems. The aim of this paper is to introduce FPA to the electromagnetics and antenna community for the optimization of linear antenna arrays. FPA is applied for the first time to linear array so as to obtain optimized antenna positions in order to achieve an array pattern with minimum side lobe level along with placement of deep nulls in desired directions. Various design examples are presented that illustrate the use of FPA for linear antenna array optimization, and subsequently the results are validated by benchmarking along with results obtained using other state-of-the-art, nature-inspired evolutionary algorithms such as particle swarm optimization, ant colony optimization and cat swarm optimization. The results suggest that in most cases, FPA outperforms the other evolutionary algorithms and at times it yields a similar performance.

Full Text Available At present, most link prediction algorithms are based on the similarity between two entities. Social network topology information is one of the main sources to design the similarity function between entities. But the existing link prediction algorithms do not apply the network topology information sufficiently. For lack of traditional link prediction algorithms, we propose two improved algorithms: CNGF algorithm based on local information and KatzGF algorithm based on global information network. For the defect of the stationary of social network, we also provide the link prediction algorithm based on nodes multiple attributes information. Finally, we verified these algorithms on DBLP data set, and the experimental results show that the performance of the improved algorithm is superior to that of the traditional link prediction algorithm.

Many linear programming models have been developed to model the logistics of bio-energy chains. These models help to determine the best set-up of bio-energy chains. Most of them use network structures built up from nodes with one or more depots, and arcs connecting these depots. Each depot is source

Optimization has become one of the important fields in Mathematics. Many problems in engineering and science can be formulated into optimization problems. They maybe have many local optima. The optimization problem with many local optima, known as multimodal optimization problem, is how to find the global solution. Several metaheuristic methods have been proposed to solve multimodal optimization problems such as Particle Swarm Optimization (PSO), Genetics Algorithm (GA), Artificial Bee Colony (ABC) algorithm, etc. The performance of the ABC algorithm is better than or similar to those of other population-based algorithms with the advantage of employing a fewer control parameters. The ABC algorithm also has the advantages of strong robustness, fast convergence and high flexibility. However, it has the disadvantages premature convergence in the later search period. The accuracy of the optimal value cannot meet the requirements sometimes. Broyden-Fletcher-Goldfarb-Shanno (BFGS) algorithm is a good iterative method for finding a local optimum. Compared with other local optimization methods, the BFGS algorithm is better. Based on the advantages of the ABC algorithm and the BFGS algorithm, this paper proposes a hybrid of the artificial bee colony algorithm and the BFGS algorithm to solve the multimodal optimization problem. The first step is that the ABC algorithm is run to find a point. In the second step is that the point obtained by the first step is used as an initial point of BFGS algorithm. The results show that the hybrid method can overcome from the basic ABC algorithm problems for almost all test function. However, if the shape of function is flat, the proposed method cannot work well.

Full Text Available Evolutionary algorithms (EAs are population-based global search methods. They have been successfully applied tomany complex optimization problems. However, EAs are frequently incapable of finding a convergence solution indefault of local search mechanisms. Memetic Algorithms (MAs are hybrid EAs that combine genetic operators withlocal search methods. With global exploration and local exploitation in search space, MAs are capable of obtainingmore high-quality solutions. On the other hand, mixed-integer hybrid differential evolution (MIHDE, as an EA-basedsearch algorithm, has been successfully applied to many mixed-integer optimization problems. In this paper, amemetic algorithm based on MIHDE is developed for solving mixed-integer optimization problems. However, most ofreal-world mixed-integer optimization problems frequently consist of equality and/or inequality constraints. In order toeffectively handle constraints, an evolutionary Lagrange method based on memetic algorithm is developed to solvethe mixed-integer constrained optimization problems. The proposed algorithm is implemented and tested on twobenchmark mixed-integer constrained optimization problems. Experimental results show that the proposed algorithmcan find better optimal solutions compared with some other search algorithms. Therefore, it implies that the proposedmemetic algorithm is a good approach to mixed-integer optimization problems.

The overall goal of this paper is to develop a general formulation for an optimal infrastructure layout design of gas pipeline distribution networks using algorithm developed from the application of two industrial engineering concepts: the traveling salesman problem (TSP) and the nearest neighbor (NN). The focus is on the ...

In this study we examine the possibility of using evolutionary optimizationalgorithms in protein-peptide docking. We present the main assumptions that reduce the docking problem to a continuous global optimization problem and provide a way of using evolutionary optimizationalgorithms. The Rosetta all-atom force field was used for structural representation and energy scoring. We describe the parallelization scheme and MPI/OpenMP realization of the considered algorithms. We demonstrate the efficiency and the performance for some algorithms which were applied to a set of benchmark tests.

Full Text Available Building and measuring trust is one of crucial aspects in e-commerce, social networking and computer security. Trust networks are widely used to formalize trust relationships and to conduct formal reasoning of trust values. Diverse trust network analysis methods have been developed so far and one of the most widely used schemes is TNA-SL (Trust Network Analysis with Subjective Logic. Recent papers claimed that TNA-SL always finds the optimal solution by producing the least uncertainty. In this paper, we present some counter-examples, which imply that TNA-SL is not an optimalalgorithm. Furthermore, we present a probabilistic algorithm in edge splitting to minimize uncertainty.

While heuristic optimization is applied in environmental applications, ad-hoc algorithm configuration is typical. We use a multi-layer sorptive barrier design problem as a benchmark for an algorithm-tuning procedure, as applied to three heuristics (genetic algorithms, simulated ...

A flexible iterative gradient algorithm for radiation therapy optimization has been developed. The algorithm is based on dose calculation using the pencil-beam description of external radiation beams in uniform and heterogeneous patients. The properties of the algorithm are described, including its ability to treat variable bounds and linear constraints, its efficiency in gradient calculation, its convergence properties and termination criteria. 116 refs

Wireless sensor network topology optimization is a highly important issue, and topology control through node selection can improve the efficiency of data forwarding, while saving energy and prolonging lifetime of the network. To address the problem of connecting a wireless sensor network to the Internet in cyber-physical systems, here we propose a geometric gateway deployment based on a competitive swarm optimizeralgorithm. The particle swarm optimization (PSO) algorithm has a continuous search feature in the solution space, which makes it suitable for finding the geometric center of gateway deployment; however, its search mechanism is limited to the individual optimum (pbest) and the population optimum (gbest); thus, it easily falls into local optima. In order to improve the particle search mechanism and enhance the search efficiency of the algorithm, we introduce a new competitive swarm optimizer (CSO) algorithm. The CSO search algorithm is based on an inter-particle competition mechanism and can effectively avoid trapping of the population falling into a local optimum. With the improvement of an adaptive opposition-based search and its ability to dynamically parameter adjustments, this algorithm can maintain the diversity of the entire swarm to solve geometric K -center gateway deployment problems. The simulation results show that this CSO algorithm has a good global explorative ability as well as convergence speed and can improve the network quality of service (QoS) level of cyber-physical systems by obtaining a minimum network coverage radius. We also find that the CSO algorithm is more stable, robust and effective in solving the problem of geometric gateway deployment as compared to the PSO or Kmedoids algorithms.

Describing a new optimizationalgorithm, the “Teaching-Learning-Based Optimization (TLBO),” in a clear and lucid style, this book maximizes reader insights into how the TLBO algorithm can be used to solve continuous and discrete optimization problems involving single or multiple objectives. As the algorithm operates on the principle of teaching and learning, where teachers influence the quality of learners’ results, the elitist version of TLBO algorithm (ETLBO) is described along with applications of the TLBO algorithm in the fields of electrical engineering, mechanical design, thermal engineering, manufacturing engineering, civil engineering, structural engineering, computer engineering, electronics engineering, physics and biotechnology. The book offers a valuable resource for scientists, engineers and practitioners involved in the development and usage of advanced optimizationalgorithms.

Quantum Monte Carlo algorithms based on a world-line representation such as the worm algorithm and the directed loop algorithm are among the most powerful numerical techniques for the simulation of non-frustrated spin models and of bosonic models. Both algorithms work in the grand-canonical ensemble and can have a winding number larger than zero. However, they retain a lot of intrinsic degrees of freedom which can be used to optimize the algorithm. We let us guide by the rigorous statements on the globally optimal form of Markov chain Monte Carlo simulations in order to devise a locally optimal formulation of the worm algorithm while incorporating ideas from the directed loop algorithm. We provide numerical examples for the soft-core Bose-Hubbard model and various spin- S models.

At the molecular level, biochemical processes are governed by random interactions between reactant molecules, and the dynamics of such systems are inherently stochastic. When the copy numbers of reactants are large, a deterministic description is adequate, but when they are small, such systems are often modeled as continuous-time Markov jump processes that can be described by the chemical master equation. Gillespie’s Stochastic Simulation Algorithm (SSA) generates exact trajectories of these systems, but the amount of computational work required for each step of the original SSA is proportional to the number of reaction channels, leading to computational complexity that scales linearly with the problem size. The original SSA is therefore inefficient for large problems, which has prompted the development of several alternative formulations with improved scaling properties. We describe an exact SSA that uses a table data structure with event time binning to achieve constant computational complexity with respect to the number of reaction channels for weakly coupled reaction networks. We present a novel adaptive binning strategy and discuss optimalalgorithm parameters. We compare the computational efficiency of the algorithm to existing methods and demonstrate excellent scaling for large problems. This method is well suited for generating exact trajectories of large weakly coupled models, including those that can be described by the reaction-diffusion master equation that arises from spatially discretized reaction-diffusion processes.

Linear Transportation Problem (LTP) is the case of constrained optimization where we want to minimize cost subject to the balance of the number of supply and the number of demand. The exact method such as northwest corner, vogel, russel, minimal cost have been applied at approaching optimal solution. In this paper, we use heurisitic like Particle Swarm Optimization (PSO) for solving linear transportation problem at any size of decision variable. In addition, we combine mutation operator of Genetic Algorithm (GA) at PSO to improve optimal solution. This method is called Particle Swarm Optimization - Genetic Algorithm (PSOGA). The simulations show that PSOGA can improve optimal solution resulted by PSO.

Full Text Available It is now well recognized that pure algorithms can be promisingly improved by hybridization with other techniques. One of the relatively new metaheuristic algorithms is Gravitational Search Algorithm (GSA which is based on the Newton laws. In this paper, to enhance the performance of GSA, a novel algorithm called “Kepler”, inspired by the astrophysics, is introduced. The Kepler algorithm is based on the principle of the first Kepler law. The hybridization of GSA and Kepler algorithm is an efficient approach to provide much stronger specialization in intensification and/or diversification. The performance of GSA–Kepler is evaluated by applying it to 14 benchmark functions with 20–1000 dimensions and the optimal approximation of linear system as a practical optimization problem. The results obtained reveal that the proposed hybrid algorithm is robust enough to optimize the benchmark functions and practical optimization problems.

Lehigh University`s Energy Research Center and the Potomac Electric Power Company have been developing software for use by plant personnel in turning a pulverized coal-fired boiler to achieve optimized combustion. The software is based on the interactions of an expert system, neural network, and a mathematical optimizationalgorithm. It uses the expert system to safely guide the plant engineer through a series of parametric boiler tests and gather a data base which characterizes boiler operation over a wide range of conditions. The neural network portion develops non-linear mapping functions between the outputs of NO{sub x}, heat rate, LOI, opacity, and the controllable boiler input parameters. These mapping functions are then analyzed by the mathematical optimizationalgorithm and the optimal boiler operating conditions are identified. This paper describes the application of the software to corner-fired boilers with either conventional or low NO{sub x} burners and overfire air.

Wireless mesh networking is a promising technology that can support numerous multimedia applications. Multimedia applications have stringent quality of service (QoS) requirements, i.e., bandwidth, delay, jitter, and packet loss ratio. Enabling such QoS-demanding applications over wireless mesh networks (WMNs) require QoS provisioning routing protocols that lead to the network resource underutilization problem. Moreover, random topology deployment leads to have some unused network resources. Therefore, resource optimization is one of the most critical design issues in multi-hop, multi-radio WMNs enabled with multimedia applications. Resource optimization has been studied extensively in the literature for wireless Ad Hoc and sensor networks, but existing studies have not considered resource underutilization issues caused by QoS provisioning routing and random topology deployment. Finding a QoS-provisioned path in wireless mesh networks is an NP complete problem. In this paper, we propose a novel Integer Linear Programming (ILP) optimization model to reconstruct the optimal connected mesh backbone topology with a minimum number of links and relay nodes which satisfies the given end-to-end QoS demands for multimedia traffic and identification of extra resources, while maintaining redundancy. We further propose a polynomial time heuristic algorithm called Link and Node Removal Considering Residual Capacity and Traffic Demands (LNR-RCTD). Simulation studies prove that our heuristic algorithm provides near-optimal results and saves about 20% of resources from being wasted by QoS provisioning routing and random topology deployment.

It is a challenging task of learning a large Bayesian network from a small data set. Most conventional structural learning approaches run into the computational as well as the statistical problems. We propose a decomposition algorithm for the structure construction without having to learn...... the complete network. The new learning algorithm firstly finds local components from the data, and then recover the complete network by joining the learned components. We show the empirical performance of the decomposition algorithm in several benchmark networks....

A method for aerodynamic shape optimization based on a genetic algorithm approach is demonstrated. The algorithm is coupled with a transonic full potential flow solver and is used to optimize the flow about transonic wings including multi-objective solutions that lead to the generation of pareto fronts. The results indicate that the genetic algorithm is easy to implement, flexible in application and extremely reliable.

Wireless communication network has such features as limited bandwidth, changeful channel and dynamic topology, etc. Ad hoc network has lots of difficulties in accessing control, bandwidth distribution, resource assign and congestion control. Therefore, a wireless packets distributing Evolutionary algorithm based on PSO (DPSO)for Ad Hoc Network is proposed. Firstly, parameters impact on performance of network are analyzed and researched to obtain network performance effective function. Secondly, the improved PSO Evolutionary Algorithm is used to solve the optimization problem from local to global in the process of network packets distributing. The simulation results show that the algorithm can ensure fairness and timeliness of network transmission, as well as improve ad hoc network resource integrated utilization efficiency.

Full Text Available The Internet in Canada is an assemblage of private and public networks. A variety of institutions and networking codes manage these networks. Conflicts exist between these parties despite their interconnection. Tensions heightened when commercial ISPs began managing traffic on their network using sophisticated routing algorithms. Concerned parties demanded legislation based on a network neutrality principle to prevent undue discrimination. While the network neutrality controversy has been addressed as a question of public policy, the controversy also includes a conflict between various codes constituting networks in Canada. The conflict between codes involve two key networking software that manifest incongruous networks. Their algorithms, the logics embedded in code, differentiate the different types of networking code. The two types of algorithms are Quality of Service and End-to-End. These algorithms treat different modalities of Internet communication differently, in part due to their deployment by different institutions. Quality of Service allows for the tiering of traffic by carriers. Commercial carriers have popularized this algorithm to promote value-added services and prevent network congestions. End-to-end algorithms, on the other hand, enforce a strict equality between modalities of communication. Peer-to-peer applications have popularized an extreme version of the end-to-algorithm, treating all nodes as equals. The popularity and growth of both these algorithms pulls the Internet in different directions, creating conflicts over its future. Through an extended review of these two algorithms and their intersection, this paper confronts how code plays a role in the network neutrality controversy.

Research on multiobjective optimization problems becomes one of the hottest topics of intelligent computation. In order to improve the search efficiency of an evolutionary algorithm and maintain the diversity of solutions, in this paper, the learning automata (LA) is first used for quantization orthogonal crossover (QOX), and a new fitness function based on decomposition is proposed to achieve these two purposes. Based on these, an orthogonal evolutionary algorithm with LA for complex multiobjective optimization problems with continuous variables is proposed. The experimental results show that in continuous states, the proposed algorithm is able to achieve accurate Pareto-optimal sets and wide Pareto-optimal fronts efficiently. Moreover, the comparison with the several existing well-known algorithms: nondominated sorting genetic algorithm II, decomposition-based multiobjective evolutionary algorithm, decomposition-based multiobjective evolutionary algorithm with an ensemble of neighborhood sizes, multiobjective optimization by LA, and multiobjective immune algorithm with nondominated neighbor-based selection, on 15 multiobjective benchmark problems, shows that the proposed algorithm is able to find more accurate and evenly distributed Pareto-optimal fronts than the compared ones.

A genetic algorithm is a powerful search technique that simulates natural evolution in order to fit a population of computational structures to the solution of an optimization problem. This technique presents several advantages over classical ones such as linear programming based techniques, often used in nuclear engineering optimization problems. However, genetic algorithms demand some extra computational cost. Nowadays, due to the fast computers available, the use of genetic algorithms has increased and its practical application has become a reality. In nuclear engineering there are many difficult optimization problems related to nuclear reactor design. Genetic algorithm is a suitable technique to face such kind of problems. This chapter presents applications of genetic algorithms for nuclear reactor core design optimization. A genetic algorithm has been designed to optimize the nuclear reactor cell parameters, such as array pitch, isotopic enrichment, dimensions and cells materials. Some advantages of this genetic algorithm implementation over a classical method based on linear programming are revealed through the application of both techniques to a simple optimization problem. In order to emphasize the suitability of genetic algorithms for design optimization, the technique was successfully applied to a more complex problem, where the classical method is not suitable. Results and comments about the applications are also presented. (orig.)

Full Text Available The wireless sensor networks are the emerging next generation sensor networks, Routing technology is the wireless sensor network communication layer of the core technology. To build reliable paths in wireless sensor networks, we can consider two ways: providing multiple paths utilizing the redundancy to assure the communication reliability or constructing transmission reliability mechanism to assure the reliability of every hop. Braid multipath algorithm and ReInforM routing algorithm are the realizations of these two mechanisms. After the analysis of these two algorithms, this paper proposes a ReInforM routing algorithm based braid multipath routing algorithm.

Currently, community detection is a hot topic. This paper, based on the self-organizing map (SOM) algorithm, introduced the idea of self-adaptation (SA) that the number of communities can be identified automatically, a novel algorithm SA-SOM of detecting communities in complex networks is proposed. Several representative real-world networks and a set of computer-generated networks by LFR-benchmark are utilized to verify the accuracy and the efficiency of this algorithm. The experimental findings demonstrate that this algorithm can identify the communities automatically, accurately and efficiently. Furthermore, this algorithm can also acquire higher values of modularity, NMI and density than the SOM algorithm does.

Background Network visualization would serve as a useful first step for analysis. However, current graph layout algorithms for biological pathways are insensitive to biologically important information, e.g. subcellular localization, biological node and graph attributes, or/and not available for large scale networks, e.g. more than 10000 elements. Results To overcome these problems, we propose the use of a biologically important graph metric, betweenness, a measure of network flow. This metric is highly correlated with many biological phenomena such as lethality and clusters. We devise a new fast parallel algorithm calculating betweenness to minimize the preprocessing cost. Using this metric, we also invent a node and edge betweenness based fast layout algorithm (BFL). BFL places the high-betweenness nodes to optimal positions and allows the low-betweenness nodes to reach suboptimal positions. Furthermore, BFL reduces the runtime by combining a sequential insertion algorim with betweenness. For a graph with n nodes, this approach reduces the expected runtime of the algorithm to O(n2) when considering edge crossings, and to O(n log n) when considering only density and edge lengths. Conclusion Our BFL algorithm is compared against fast graph layout algorithms and approaches requiring intensive optimizations. For gene networks, we show that our algorithm is faster than all layout algorithms tested while providing readability on par with intensive optimizationalgorithms. We achieve a 1.4 second runtime for a graph with 4000 nodes and 12000 edges on a standard desktop computer. PMID:19146673

Highlights: • Four hybrid algorithms are proposed for the wind speed decomposition. • Adaboost algorithm is adopted to provide a hybrid training framework. • MLP neural networks are built to do the forecasting computation. • Four important network training algorithms are included in the MLP networks. • All the proposed hybrid algorithms are suitable for the wind speed predictions. - Abstract: The technology of wind speed prediction is important to guarantee the safety of wind power utilization. In this paper, four different hybrid methods are proposed for the high-precision multi-step wind speed predictions based on the Adaboost (Adaptive Boosting) algorithm and the MLP (Multilayer Perceptron) neural networks. In the hybrid Adaboost–MLP forecasting architecture, four important algorithms are adopted for the training and modeling of the MLP neural networks, including GD-ALR-BP algorithm, GDM-ALR-BP algorithm, CG-BP-FR algorithm and BFGS algorithm. The aim of the study is to investigate the promoted forecasting percentages of the MLP neural networks by the Adaboost algorithm’ optimization under various training algorithms. The hybrid models in the performance comparison include Adaboost–GD-ALR-BP–MLP, Adaboost–GDM-ALR-BP–MLP, Adaboost–CG-BP-FR–MLP, Adaboost–BFGS–MLP, GD-ALR-BP–MLP, GDM-ALR-BP–MLP, CG-BP-FR–MLP and BFGS–MLP. Two experimental results show that: (1) the proposed hybrid Adaboost–MLP forecasting architecture is effective for the wind speed predictions; (2) the Adaboost algorithm has promoted the forecasting performance of the MLP neural networks considerably; (3) among the proposed Adaboost–MLP forecasting models, the Adaboost–CG-BP-FR–MLP model has the best performance; and (4) the improved percentages of the MLP neural networks by the Adaboost algorithm decrease step by step with the following sequence of training algorithms as: GD-ALR-BP, GDM-ALR-BP, CG-BP-FR and BFGS

The network geometry strongly influences the performance of the distributed system, i.e., the coverage capability, measurement accuracy and overall cost. Therefore the network placement optimization represents an urgent issue in the distributed measurement, even in large-scale metrology. This paper presents an effective computer-assisted network placement optimization procedure for the large-scale distributed system and illustrates it with the example of the multi-tracker system. To get an optimal placement, the coverage capability and the coordinate uncertainty of the network are quantified. Then a placement optimization objective function is developed in terms of coverage capabilities, measurement accuracy and overall cost. And a novel grid-based encoding approach for Genetic algorithm is proposed. So the network placement is optimized by a global rough search and a local detailed search. Its obvious advantage is that there is no need for a specific initial placement. At last, a specific application illustrates this placement optimization procedure can simulate the measurement results of a specific network and design the optimal placement efficiently.

We give an overview of ISINA: INTEGRAL Source Identification NetworkAlgorithm. This machine learning algorithm, using random forests, is applied to the IBIS/ISGRI data set in order to ease the production of unbiased future soft gamma-ray source catalogues. First, we introduce the data set and the problems encountered when dealing with images obtained using the coded mask technique. The initial step of source candidate searching is introduced and an initial candidate list is created. A description of the feature extraction on the initial candidate list is then performed together with feature merging for these candidates. Three training and testing sets are created in order to deal with the diverse time-scales encountered when dealing with the gamma-ray sky. Three independent random forests are built: one dealing with faint persistent source recognition, one dealing with strong persistent sources and a final one dealing with transients. For the latter, a new transient detection technique is introduced and described: the transient matrix. Finally the performance of the network is assessed and discussed using the testing set and some illustrative source examples. Based on observations with INTEGRAL, an ESA project with instruments and science data centre funded by ESA member states (especially the PI countries: Denmark, France, Germany, Italy, Spain), Czech Republic and Poland, and the participation of Russia and the USA. E-mail: simo@astro.soton.ac.uk

Full Text Available In this paper, a novel multiobjective optimization model is presented for selecting optimal locations in the water distribution network (WDN with the aim of installing pressure loggers. The pressure data collected at optimal locations will be used later on in the calibration of the proposed WDN model. Objective functions consist of maximization of calibrated model prediction accuracy and minimization of the total cost for sampling design. In order to decrease the model run time, an optimization model has been developed using multiobjective genetic algorithm and adaptive neural network (MOGA-ANN. Neural networks (NNs are initially trained after a number of initial GA generations and periodically retrained and updated after generation of a specified number of full model-analyzed solutions. Trained NNs are replaced with the fitness evaluation of some chromosomes within the GA progress. Using cache prevents objective function evaluation of repetitive chromosomes within GA. Optimal solutions are obtained through pareto-optimal front with respect to the two objective functions. Results show that jointing NNs in MOGA for approximating portions of chromosomes’ fitness in each generation leads to considerable savings in model run time and can be promising for reducing run-time in optimization models with significant computational effort.

Pressurizer is one of key components in nuclear power system. It's important to control the dimension in the design of pressurizer through optimization techniques. In this work, a mathematic model of a vertical electric heating pressurizer was established. A new Genetic-Simplex Algorithm (GSA) that combines genetic algorithm and simplex algorithm was developed to enhance the searching ability, and the comparison among modified and original algorithms is conducted by calculating the benchmark function. Furthermore, the optimization design of pressurizer, taking minimization of volume and net weight as objectives, was carried out considering thermal-hydraulic and geometric constraints through GSA. The results indicate that the mathematical model is agreeable for the pressurizer and the new algorithm is more effective than the traditional genetic algorithm. The optimization design shows obvious validity and can provide guidance for real engineering design.

Full Text Available This paper investigates an evolutionary-based designing system for automated sizing of analog integrated circuits (ICs. Two evolutionary algorithms, genetic algorithm and PSO (Parswal particle swarm optimizationalgorithm, are proposed to design analog ICs with practical user-defined specifications. On the basis of the combination of HSPICE and MATLAB, the system links circuit performances, evaluated through specific electrical simulation, to the optimization system in the MATLAB environment, for the selected topology. The system has been tested by typical and hard-to-design cases, such as complex analog blocks with stringent design requirements. The results show that the design specifications are closely met. Comparisons with available methods like genetic algorithms show that the proposed algorithm offers important advantages in terms of optimization quality and robustness. Moreover, the algorithm is shown to be efficient.

Pressurizer is one of key components in nuclear power system. It's important to control the dimension in the design of pressurizer through optimization techniques. In this work, a mathematic model of a vertical electric heating pressurizer was established. A new Genetic-Simplex Algorithm (GSA) that combines genetic algorithm and simplex algorithm was developed to enhance the searching ability, and the comparison among modified and original algorithms is conducted by calculating the benchmark function. Furthermore, the optimization design of pressurizer, taking minimization of volume and net weight as objectives, was carried out considering thermal-hydraulic and geometric constraints through GSA. The results indicate that the mathematical model is agreeable for the pressurizer and the new algorithm is more effective than the traditional genetic algorithm. The optimization design shows obvious validity and can provide guidance for real engineering design

In this work, we present an extension of the forward-reverse algorithm by Bayer and Schoenmakers [2] to the context of stochastic reaction networks (SRNs). We then apply this bridge-generation technique to the statistical inference problem of approximating the reaction coefficients based on discretely observed data. To this end, we introduce a two-phase iterative inference method in which we solve a set of deterministic optimization problems where the SRNs are replaced by the classical ODE rates; then, during the second phase, the Monte Carlo version of the EM algorithm is applied starting from the output of the previous phase. Starting from a set of over-dispersed seeds, the output of our two-phase method is a cluster of maximum likelihood estimates obtained by using convergence assessment techniques from the theory of Markov chain Monte Carlo.

This book presents the latest findings on stochastic dynamic programming models and on solving optimal control problems in networks. It includes the authors' new findings on determining the optimal solution of discrete optimal control problems in networks and on solving game variants of Markov decision problems in the context of computational networks. First, the book studies the finite state space of Markov processes and reviews the existing methods and algorithms for determining the main characteristics in Markov chains, before proposing new approaches based on dynamic programming and combinatorial methods. Chapter two is dedicated to infinite horizon stochastic discrete optimal control models and Markov decision problems with average and expected total discounted optimization criteria, while Chapter three develops a special game-theoretical approach to Markov decision processes and stochastic discrete optimal control problems. In closing, the book's final chapter is devoted to finite horizon stochastic con...

Correct inference of genetic regulations inside a cell from the biological database like time series microarray data is one of the greatest challenges in post genomic era for biologists and researchers. Recurrent Neural Network (RNN) is one of the most popular and simple approach to model the dynamics as well as to infer correct dependencies among genes. Inspired by the behavior of social elephants, we propose a new metaheuristic namely Elephant Swarm Water Search Algorithm (ESWSA) to infer Gene Regulatory Network (GRN). This algorithm is mainly based on the water search strategy of intelligent and social elephants during drought, utilizing the different types of communication techniques. Initially, the algorithm is tested against benchmark small and medium scale artificial genetic networks without and with presence of different noise levels and the efficiency was observed in term of parametric error, minimum fitness value, execution time, accuracy of prediction of true regulation, etc. Next, the proposed algorithm is tested against the real time gene expression data of Escherichia Coli SOS Network and results were also compared with others state of the art optimization methods. The experimental results suggest that ESWSA is very efficient for GRN inference problem and performs better than other methods in many ways.

This paper presents a probabilistic computational framework for the Pareto optimization of the preventive maintenance applications to bridges of a highway transportation network. The bridge characteristics are represented by their uncertain reliability index profiles. The in/out of service states of the bridges are simulated taking into account their correlation structure. Multi-objective Genetic Algorithms have been chosen as numerical tool for the solution of the optimization problem. The design variables of the optimization are the preventive maintenance schedules of all the bridges of the network. The two conflicting objectives are the minimization of the total present maintenance cost and the maximization of the network performance indicator. The final result is the Pareto front of optimal solutions among which the managers should chose, depending on engineering and economical factors. A numerical example illustrates the application of the proposed approach.

A cognitive radio (CR) network with energy harvesting (EH) is considered to improve both spectrum efficiency and energy efficiency. A hidden Markov model (HMM) is used to characterize the imperfect spectrum sensing process. In order to maximize the whole satisfaction degree (WSD) of the cognitive radio network, a tradeoff between the average throughput of the secondary user (SU) and the interference to the primary user (PU) is analyzed. We formulate the satisfaction degree optimization problem as a mixed integer nonlinear programming (MINLP) problem. The satisfaction degree optimization problem is solved by using differential evolution (DE) algorithm. The proposed optimization problem allows the network to adaptively achieve the optimal solution based on its required quality of service (Qos). Numerical results are given to verify our analysis. Project supported by the National Natural Science Foundation of China (Grant No. 61301179), the Doctorial Programs Foundation of the Ministry of Education of China (Grant No. 20110203110011), and the 111 Project (Grant No. B08038).

In this paper, we propose a new adaptive unified differential evolution algorithm for single-objective global optimization. Instead of the multiple mutation strate- gies proposed in conventional differential evolution algorithms, this algorithm employs a single equation unifying multiple strategies into one expression. It has the virtue of mathematical simplicity and also provides users the flexibility for broader exploration of the space of mutation operators. By making all control parameters in the proposed algorithm self-adaptively evolve during the process of optimization, it frees the application users from the burden of choosing appro- priate control parameters and also improves the performance of the algorithm. In numerical tests using thirteen basic unimodal and multimodal functions, the proposed adaptive unified algorithm shows promising performance in compari- son to several conventional differential evolution algorithms.

Full Text Available Two new PRP conjugate Algorithms are proposed in this paper based on two modified PRP conjugate gradient methods: the first algorithm is proposed for solving unconstrained optimization problems, and the second algorithm is proposed for solving nonlinear equations. The first method contains two aspects of information: function value and gradient value. The two methods both possess some good properties, as follows: 1 βk ≥ 0 2 the search direction has the trust region property without the use of any line search method 3 the search direction has sufficient descent property without the use of any line search method. Under some suitable conditions, we establish the global convergence of the two algorithms. We conduct numerical experiments to evaluate our algorithms. The numerical results indicate that the first algorithm is effective and competitive for solving unconstrained optimization problems and that the second algorithm is effective for solving large-scale nonlinear equations.

activities with the Autonomous Component Carrier Selection (ACCS) algorithm, a distributed solution for interference management among small neighboring cells. A preliminary evaluation of the algorithm performance is provided considering its live execution on a software defined radio network testbed....... The obtained experimental results confirm the performance trends obtained from prior simulation studies. The analysis in dynamic environment conditions also allowed identifying the utilization of static thresholds in the decision making process, as a critical aspect for the optimization of network capacity....

This paper proposes a novel optimization approach for the least cost design of looped water distribution systems (WDSs). Three distinct steps are involved in the proposed optimization approach. In the first step, the shortest-distance tree within the looped network is identified using the Dijkstra graph theory algorithm, for which an extension is proposed to find the shortest-distance tree for multisource WDSs. In the second step, a nonlinear programming (NLP) solver is employed to optimize the pipe diameters for the shortest-distance tree (chords of the shortest-distance tree are allocated the minimum allowable pipe sizes). Finally, in the third step, the original looped water network is optimized using a differential evolution (DE) algorithm seeded with diameters in the proximity of the continuous pipe sizes obtained in step two. As such, the proposed optimization approach combines the traditional deterministic optimization technique of NLP with the emerging evolutionary algorithm DE via the proposed network decomposition. The proposed methodology has been tested on four looped WDSs with the number of decision variables ranging from 21 to 454. Results obtained show the proposed approach is able to find optimal solutions with significantly less computational effort than other optimization techniques.

This paper presents the Particle Swarm Optimization (PSO) algorithm to improve the quality of the power supply in a microgrid. This algorithm is proposed for a real-time selftuning method that used in a power controller for an inverter based Distributed Generation (DG) unit. In such system, the voltage and frequency are the main control objectives, particularly when the microgrid is islanded or during load change. In this work, the PSO algorithm is implemented to find the optimal controller parameters to satisfy the control objectives. The results show high performance of the applied PSO algorithm of regulating the microgrid voltage and frequency.

In this paper, an algorithm to choose a good partition in bipartite networks has been proposed. Bipartite networks have more theoretical significance and broader prospect of application. In view of distinctive structure of bipartite networks, in our method, two parameters are defined to show the relationships between the same type nodes and heterogeneous nodes respectively. Moreover, our algorithm employs a new method of finding and expanding the core communities in bipartite networks. Two kinds of nodes are handled separately and merged, and then the sub-communities are obtained. After that, objective communities will be found according to the merging rule. The proposed algorithm has been simulated in real-world networks and artificial networks, and the result verifies the accuracy and reliability of the parameters on intimacy for our algorithm. Eventually, comparisons with similar algorithms depict that the proposed algorithm has better performance.

It has been widely demonstrated that the structure of a network is a major factor that affects its traffic dynamics. In this work, we try to identify the optimal topologies for maximizing the network transmission capacity, as well as to build a clear relationship between structural features of a network and the transmission performance in terms of traffic delivery. We propose an approach for designing optimalnetwork topologies against traffic congestion by link rewiring and apply them on the Barabási-Albert scale-free, static scale-free and Internet Autonomous System-level networks. Furthermore, we analyze the optimizednetworks using complex network parameters that characterize the structure of networks, and our simulation results suggest that an optimalnetwork for traffic transmission is more likely to have a core-periphery structure. However, assortative mixing and the rich-club phenomenon may have negative impacts on network performance. Based on the observations of the optimizednetworks, we propose an efficient method to improve the transmission capacity of large-scale networks.

A typical modern optimization technique is usually either heuristic or metaheuristic. This technique has managed to solve some optimization problems in the research area of science, engineering, and industry. However, implementation strategy of metaheuristic for accuracy improvement on convolution neural networks (CNN), a famous deep learning method, is still rarely investigated. Deep learning relates to a type of machine learning technique, where its aim is to move closer to the goal of artificial intelligence of creating a machine that could successfully perform any intellectual tasks that can be carried out by a human. In this paper, we propose the implementation strategy of three popular metaheuristic approaches, that is, simulated annealing, differential evolution, and harmony search, to optimize CNN. The performances of these metaheuristic methods in optimizing CNN on classifying MNIST and CIFAR dataset were evaluated and compared. Furthermore, the proposed methods are also compared with the original CNN. Although the proposed methods show an increase in the computation time, their accuracy has also been improved (up to 7.14 percent).

Full Text Available A typical modern optimization technique is usually either heuristic or metaheuristic. This technique has managed to solve some optimization problems in the research area of science, engineering, and industry. However, implementation strategy of metaheuristic for accuracy improvement on convolution neural networks (CNN, a famous deep learning method, is still rarely investigated. Deep learning relates to a type of machine learning technique, where its aim is to move closer to the goal of artificial intelligence of creating a machine that could successfully perform any intellectual tasks that can be carried out by a human. In this paper, we propose the implementation strategy of three popular metaheuristic approaches, that is, simulated annealing, differential evolution, and harmony search, to optimize CNN. The performances of these metaheuristic methods in optimizing CNN on classifying MNIST and CIFAR dataset were evaluated and compared. Furthermore, the proposed methods are also compared with the original CNN. Although the proposed methods show an increase in the computation time, their accuracy has also been improved (up to 7.14 percent.

Multiobjective design optimization problems require multiobjective optimization techniques to solve, and it is often very challenging to obtain high-quality Pareto fronts accurately. In this article, the recently developed flower pollination algorithm (FPA) is extended to solve multiobjective optimization problems. The proposed method is used to solve a set of multiobjective test functions and two bi-objective design benchmarks, and a comparison of the proposed algorithm with other algorithms has been made, which shows that the FPA is efficient with a good convergence rate. Finally, the importance for further parametric studies and theoretical analysis is highlighted and discussed.

Studies on wear properties of Aluminium based hybrid nano composite materials, processed through powder metallurgy technique, are reported in the present study. Silicon Carbide nano particles and E-glass fibre are reinforced in pure aluminium matrix to fabricate hybrid nano composite material samples. Pin-on-Disc wear testing equipment is used to evaluate dry sliding wear properties of the composite samples. The tests were conducted following the Taguchi’s Design of Experiments method. Signal-to-Noise ratio analysis and Analysis of Variance are carried out on the test data to find out the influence of test parameters on the wear rate. Scanning Electron Microscopic analysis and Energy Dispersive x-ray analysis are conducted on the worn surfaces to find out the wear mechanisms responsible for wear of the composites. Multiple linear regression analysis and Genetic Algorithm techniques are employed for optimization of wear test parameters to yield minimum wear of the composite samples. Finally, a wear model is built by the application of Artificial Neural Networks to predict the wear rate of the composite material, under different testing conditions. The predicted values of wear rate are found to be very close to the experimental values with a deviation in the range of 0.15% to 8.09%.

Full Text Available Power generation industry is the key industry of carbon dioxide (CO2 emission in China. Assessing its future CO2 emissions is of great significance to the formulation and implementation of energy saving and emission reduction policies. Based on the Stochastic Impacts by Regression on Population, Affluence and Technology model (STIRPAT, the influencing factors analysis model of CO2 emission of power generation industry is established. The ridge regression (RR method is used to estimate the historical data. In addition, a wavelet neural network (WNN prediction model based on Cuckoo Search algorithmoptimized by Gauss (GCS is put forward to predict the factors in the STIRPAT model. Then, the predicted values are substituted into the regression model, and the CO2 emission estimation values of the power generation industry in China are obtained. It’s concluded that population, per capita Gross Domestic Product (GDP, standard coal consumption and thermal power specific gravity are the key factors affecting the CO2 emission from the power generation industry. Besides, the GCS-WNN prediction model has higher prediction accuracy, comparing with other models. Moreover, with the development of science and technology in the future, the CO2 emission growth in the power generation industry will gradually slow down according to the prediction results.

Reduced graphene oxide-supported nanoscale zero-valent iron (nZVI/rGO) composites were synthesized in the present study by chemical deposition method and were then characterized by various methods, such as Fourier-transform infrared spectroscopy (FTIR) and X-ray photoelectron spectroscopy (XPS). The nZVI/rGO composites prepared were utilized for Cd(II) removal from aqueous solutions in batch mode at different initial Cd(II) concentrations, initial pH values, contact times, and operating temperatures. Response surface methodology (RSM) and artificial neural network hybridized with genetic algorithm (ANN-GA) were used for modeling the removal efficiency of Cd(II) and optimizing the four removal process variables. The average values of prediction errors for the RSM and ANN-GA models were 6.47% and 1.08%. Although both models were proven to be reliable in terms of predicting the removal efficiency of Cd(II), the ANN-GA model was found to be more accurate than the RSM model. In addition, experimental data were fitted to the Langmuir, Freundlich, and Dubinin-Radushkevich (D-R) isotherms. It was found that the Cd(II) adsorption was best fitted to the Langmuir isotherm. Examination on thermodynamic parameters revealed that the removal process was spontaneous and exothermic in nature. Furthermore, the pseudo-second-order model can better describe the kinetics of Cd(II) removal with a good R² value than the pseudo-first-order model.

Full Text Available Reduced graphene oxide-supported Fe3O4 (Fe3O4/rGO composites were applied in this study to remove low-concentration mercury from aqueous solutions with the aid of an artificial neural network (ANN modeling and genetic algorithm (GA optimization. The Fe3O4/rGO composites were prepared by the solvothermal method and characterized by X-ray diffraction (XRD, transmission electron microscopy (TEM, atomic force microscopy (AFM, N2-sorption, X-ray photoelectron spectroscopy (XPS, Fourier transform infrared spectroscopy (FTIR and superconduction quantum interference device (SQUID. Response surface methodology (RSM and ANN were employed to model the effects of different operating conditions (temperature, initial pH, initial Hg ion concentration and contact time on the removal of the low-concentration mercury from aqueous solutions by the Fe3O4/rGO composites. The ANN-GA model results (with a prediction error below 5% show better agreement with the experimental data than the RSM model results (with a prediction error below 10%. The removal process of the low-concentration mercury obeyed the Freudlich isotherm and the pseudo-second-order kinetic model. In addition, a regeneration experiment of the Fe3O4/rGO composites demonstrated that these composites can be reused for the removal of low-concentration mercury from aqueous solutions.

This paper investigates the problem of tracking highly dynamic targets in real-time. The advantages and disadvantages of various control and interpolation techniques are discussed. Algorithms appropriate for real- time tracking are compared with algorithms that are common in applications with a priori knowledge of the trajectory. A unique command interpolation technique combined with a state feedback control structure, implemented on a commercial DSP controller, demonstrates a significant performance improvement.

Data acquisition is the foundation of soft sensor and data fusion. Distributed data acquisition and its synchronization are the important technologies to ensure the accuracy of soft sensors. As a research topic in bionic science, the firefly-inspired algorithm has attracted widespread attention as a new synchronization method. Aiming at reducing the design difficulty of firefly-inspired synchronization algorithms for Wireless Sensor Networks (WSNs) with complex topologies, this paper presents a firefly-inspired synchronization algorithm based on a multiscale discrete phase model that can optimize the performance tradeoff between the network scalability and synchronization capability in a complex wireless sensor network. The synchronization process can be regarded as a Markov state transition, which ensures the stability of this algorithm. Compared with the Miroll and Steven model and Reachback Firefly Algorithm, the proposed algorithm obtains better stability and performance. Finally, its practicality has been experimentally confirmed using 30 nodes in a real multi-hop topology with low quality links.

Full Text Available Abstract Background A reverse engineering of gene regulatory network with large number of genes and limited number of experimental data points is a computationally challenging task. In particular, reverse engineering using linear systems is an underdetermined and ill conditioned problem, i.e. the amount of microarray data is limited and the solution is very sensitive to noise in the data. Therefore, the reverse engineering of gene regulatory networks with large number of genes and limited number of data points requires rigorous optimizationalgorithm. Results This study presents a novel algorithm for reverse engineering with linear systems. The proposed algorithm is a combination of the orthogonal least squares, second order derivative for network pruning, and Bayesian model comparison. In this study, the entire network is decomposed into a set of small networks that are defined as unit networks. The algorithm provides each unit network with P(D|Hi, which is used as confidence level. The unit network with higher P(D|Hi has a higher confidence such that the unit network is correctly elucidated. Thus, the proposed algorithm is able to locate true positive interactions using P(D|Hi, which is a unique property of the proposed algorithm. The algorithm is evaluated with synthetic and Saccharomyces cerevisiae expression data using the dynamic Bayesian network. With synthetic data, it is shown that the performance of the algorithm depends on the number of genes, noise level, and the number of data points. With Yeast expression data, it is shown that there is remarkable number of known physical or genetic events among all interactions elucidated by the proposed algorithm. The performance of the algorithm is compared with Sparse Bayesian Learning algorithm using both synthetic and Saccharomyces cerevisiae expression data sets. The comparison experiments show that the algorithm produces sparser solutions with less false positives than Sparse Bayesian

Distributed energy resources (DERs) are increasingly penetrating the energy system which is driven by climate and sustainability goals. These technologies are mostly connected to low- voltage electrical networks and change the demand and supply situation in these networks. This can cause critical network states. Network topologies vary significantly and depend on several conditions including geography, historical development, network design or number of network connections. In the past, only some of these aspects were taken into account when estimating the network investment needs for Germany on the low-voltage level. Typically, fixed network topologies are examined or a Monte Carlo approach is used to quantify the investment needs at this voltage level. Recent research has revealed that DERs differ substantially between rural, suburban and urban regions. The low-voltage network topologies have different design concepts in these regions, so that different network topologies have to be considered when assessing the need for network extensions and investments due to DERs. An extension algorithm is needed to calculate network extensions and investment needs for the different typologies of generic low-voltage networks. We therefore present a new algorithm, which is capable of calculating the extension for generic low-voltage networks of any given topology based on voltage range deviations and thermal overloads. The algorithm requires information about line and cable lengths, their topology and the network state only. We test the algorithm on a radial, a loop, and a heavily meshed network. Here we show that the algorithm functions for electrical networks with these topologies. We found that the algorithm is able to extend different networks efficiently by placing cables between network nodes. The main value of the algorithm is that it does not require any information about routes for additional cables or positions for additional substations when it comes to estimating

The training of product neural networks using genetic algorithms is discussed. Two unusual neural network techniques are combined; product units are employed instead of the traditional summing units and genetic algorithms train the network rather than backpropagation. As an example, a neural netork is trained to calculate the optimum width of transistors in a CMOS switch. It is shown how local minima affect the performance of a genetic algorithm, and one method of overcoming this is presented.

Considering the influence of various factors, such as speed, angle, depth of water, weight, and water flow, on the underwater rescue robot, a method based on neural network is proposed. According to the characteristics of Elman and Jordan neural network, a new dynamic neural network is constructed. The network can be used to remember the state of the hidden layer and increase the feedback of the output node. The improved Jordan network is optimized by chaos particle swarm optimizationalgorithm. The optimized neural network is applied to identify the dynamic model of the underwater rescue robot. The simulation results show that the neural network has good convergence speed and accuracy.

Several issues in the design and optimization of the next generation optical access network (NG-OAN) are presented. The noise, the distortion and the fiber optic nonlinearities are considered to optimize the video distribution link in a passive optical network (PON). A discussion of the effect...

The article deals with the issue of highways network design. Studies show that the main requirement from road transport for the road network is to ensure the realization of all the transport links served by it, with the least possible cost. The goal of optimizing the network of highways is to increase the efficiency of transport. It is necessary to take into account a large number of factors that make it difficult to quantify and qualify their impact on the road network. In this paper, we propose building an optimal variant for locating the road network on the basis of a mathematical model. The article defines the criteria for optimality and objective functions that reflect the requirements for the road network. The most fully satisfying condition for optimality is the minimization of road and transport costs. We adopted this indicator as a criterion of optimality in the economic-mathematical model of a network of highways. Studies have shown that each offset point in the optimal binding road network is associated with all other corresponding points in the directions providing the least financial costs necessary to move passengers and cargo from this point to the other corresponding points. The article presents general principles for constructing an optimalnetwork of roads.

Full Text Available Harmony search algorithm (HS is a new metaheuristic algorithm which is inspired by a process involving musical improvisation. HS is a stochastic optimization technique that is similar to genetic algorithms (GAs and particle swarm optimizers (PSOs. It has been widely applied in order to solve many complex optimization problems, including continuous and discrete problems, such as structure design, and function optimization. A cooperative harmony search algorithm (CHS is developed in this paper, with cooperative behavior being employed as a significant improvement to the performance of the original algorithm. Standard HS just uses one harmony memory and all the variables of the object function are improvised within the harmony memory, while the proposed algorithm CHS uses multiple harmony memories, so that each harmony memory can optimize different components of the solution vector. The CHS was then applied to function optimization problems. The results of the experiment show that CHS is capable of finding better solutions when compared to HS and a number of other algorithms, especially in high-dimensional problems.

Full Text Available Genetic algorithm (GA is utilized to select most suitable Distributed Generator (DG technology for optimal operation of power system as well as determine the optimal location and size of the DG to minimize power loss on the network. Three classes of DG technologies, synchronous generators, asynchronous generators, and induction generators, are considered and included as part of the variables for the optimization problem. IEEE 14-bus network is used to test the applicability of the algorithm. The result reveals that the developed algorithm is able to successfully select the most suitable DG technology and optimally size and place the DGs to minimize power loss in the network. Furthermore, optimum multiple placement of DG is considered to see the possible impact on power loss in the network. The result reveals that multiple placements can further reduce the power loss in the network.

Many decision problems in Operations Research are defined on temporal networks, that is, workflows of time-consuming tasks whose processing order is constrained by precedence relations. For example, temporal networks are used to model projects, computer applications, digital circuits and production processes. Optimization problems arise in temporal networks when a decision maker wishes to determine a temporal arrangement of the tasks and/or a resource assignment that optimizes some network characteristic (e.g. the time required to complete all tasks). The parameters of these optimization probl

Abstract?In this paper, we propose a new unified differential evolution (uDE) algorithm for single objective global optimization. Instead of selecting among multiple mutation strategies as in the conventional differential evolution algorithm, this algorithm employs a single equation as the mutation strategy. It has the virtue of mathematical simplicity and also provides users the flexbility for broader exploration of different mutation strategies. Numerical tests using twelve basic unimodal and multimodal functions show promising performance of the proposed algorithm in comparison to convential differential evolution algorithms.

Full Text Available This paper presents a new bio-inspired heuristic optimizationalgorithm called the Wolf Search Algorithm (WSA for solving the multi-objective reactive power dispatch problem. Wolf Search algorithm is a new bio – inspired heuristic algorithm which based on wolf preying behaviour. The way wolves search for food and survive by avoiding their enemies has been imitated to formulate the algorithm for solving the reactive power dispatches. And the speciality of wolf is possessing both individual local searching ability and autonomous flocking movement and this special property has been utilized to formulate the search algorithm .The proposed (WSA algorithm has been tested on standard IEEE 30 bus test system and simulation results shows clearly about the good performance of the proposed algorithm .

Recently much attention has been devoted to the optimization of transportation networks in a given geographic area. One assumes the distributions of population and of services/workplaces (i.e. the network's sources and sinks) are known, as well as the costs of movement with/without the network, and the cost of constructing/maintaining it. Both the long-term optimization and the short-term, "who goes where" optimization are considered. These models can also be adapted for the optimization of other types of networks, such as telecommunications, pipeline or drainage networks. In the monograph we study the most general problem settings, namely, when neither the shape nor even the topology of the network to be constructed is known a priori.

Full Text Available The security incidents ion networks are sudden and uncertain, it is very hard to precisely predict the network security situation by traditional methods. In order to improve the prediction accuracy of the network security situation, we build a network security situation prediction model based on Wavelet Neural Network (WNN with optimized parameters by the Improved Niche Genetic Algorithm (INGA. The proposed model adopts WNN which has strong nonlinear ability and fault-tolerance performance. Also, the parameters for WNN are optimized through the adaptive genetic algorithm (GA so that WNN searches more effectively. Considering the problem that the adaptive GA converges slowly and easily turns to the premature problem, we introduce a novel niche technology with a dynamic fuzzy clustering and elimination mechanism to solve the premature convergence of the GA. Our final simulation results show that the proposed INGA-WNN prediction model is more reliable and effective, and it achieves faster convergence-speed and higher prediction accuracy than the Genetic Algorithm-Wavelet Neural Network (GA-WNN, Genetic Algorithm-Back Propagation Neural Network (GA-BPNN and WNN.

Preliminary results are presented on the development of an adaptive neural network based control algorithm to enhance aircraft engine performance. This work builds upon a previous National Aeronautics and Space Administration (NASA) effort known as Performance Seeking Control (PSC). PSC is an adaptive control algorithm which contains a model of the aircraft's propulsion system which is updated on-line to match the operation of the aircraft's actual propulsion system. Information from the on-line model is used to adapt the control system during flight to allow optimal operation of the aircraft's propulsion system (inlet, engine, and nozzle) to improve aircraft engine performance without compromising reliability or operability. Performance Seeking Control has been shown to yield reductions in fuel flow, increases in thrust, and reductions in engine fan turbine inlet temperature. The neural network based adaptive control, like PSC, will contain a model of the propulsion system which will be used to calculate optimal control commands on-line. Hopes are that it will be able to provide some additional benefits above and beyond those of PSC. The PSC algorithm is computationally intensive, it is valid only at near steady-state flight conditions, and it has no way to adapt or learn on-line. These issues are being addressed in the development of the optimal neural controller. Specialized neural network processing hardware is being developed to run the software, the algorithm will be valid at steady-state and transient conditions, and will take advantage of the on-line learning capability of neural networks. Future plans include testing the neural network software and hardware prototype against an aircraft engine simulation. In this paper, the proposed neural network software and hardware is described and preliminary neural network training results are presented.

Aiming at the disadvantage of conventional optimization method for ball mill pulverizing system, a novel approach based on RBF neural network and genetic algorithm was proposed in the present paper. Firstly, the experiments and measurement for fill level based on vibration signals of mill shell was introduced. Then, main factors which affected the power consumption of ball mill pulverizing system were analyzed, and the input variables of RBF neural network were determined. RBF neural network was used to map the complex non-linear relationship between the electric consumption and process parameters and the non-linear model of power consumption was built. Finally, the model was optimized by genetic algorithm and the optimal work conditions of ball mill pulverizing system were determined. The results demonstrate that the method is reliable and practical, and can reduce the electric consumption obviously and effectively.

Full Text Available The relation between network structure and dynamics is determinant for the behavior of complex systems in numerous domains. An important long-standing problem concerns the properties of the networks that optimize the dynamics with respect to a given performance measure. Here, we show that such optimization can lead to sensitive dependence of the dynamics on the structure of the network. Specifically, using diffusively coupled systems as examples, we demonstrate that the stability of a dynamical state can exhibit sensitivity to unweighted structural perturbations (i.e., link removals and node additions for undirected optimalnetworks and to weighted perturbations (i.e., small changes in link weights for directed optimalnetworks. As mechanisms underlying this sensitivity, we identify discontinuous transitions occurring in the complement of undirected optimalnetworks and the prevalence of eigenvector degeneracy in directed optimalnetworks. These findings establish a unified characterization of networksoptimized for dynamical stability, which we illustrate using Turing instability in activator-inhibitor systems, synchronization in power-grid networks, network diffusion, and several other network processes. Our results suggest that the network structure of a complex system operating near an optimum can potentially be fine-tuned for a significantly enhanced stability compared to what one might expect from simple extrapolation. On the other hand, they also suggest constraints on how close to the optimum the system can be in practice. Finally, the results have potential implications for biophysical networks, which have evolved under the competing pressures of optimizing fitness while remaining robust against perturbations.

Recent work has demonstrated the ability of neural-network-based controllers to optimize and control machines with complex, non-linear, relatively unknown control spaces. We present a brief overview of neural networks via a taxonomy illustrating some capabilities of different kinds of neural networks. We present some successful control examples, particularly the optimization and control of a small-angle negative ion source.

In this paper, we discuss the portfolio optimization problem with real-world constraints under the assumption that the returns of risky assets are fuzzy numbers. A new possibilistic mean-semiabsolute deviation model is proposed, in which transaction costs, cardinality and quantity constraints are considered. Due to such constraints the proposed model becomes a mixed integer nonlinear programming problem and traditional optimization methods fail to find the optimal solution efficiently. Thus, a modified artificial bee colony (MABC) algorithm is developed to solve the corresponding optimization problem. Finally, a numerical example is given to illustrate the effectiveness of the proposed model and the corresponding algorithm.

Full Text Available A hybrid trajectory optimization method consisting of Gauss pseudospectral method (GPM and natural computation algorithm has been developed and utilized to solve multiphase return trajectory optimization problem, where a phase is defined as a subinterval in which the right-hand side of the differential equation is continuous. GPM converts the optimal control problem to a nonlinear programming problem (NLP, which helps to improve calculation accuracy and speed of natural computation algorithm. Through numerical simulations, it is found that the multiphase optimal control problem could be solved perfectly.

Firefly Algorithm (FA) is a metaheuristic algorithm that is inspired by the flashing behavior of fireflies and the phenomenon of bioluminescent communication and the algorithm is used to optimize the machining parameters (feed rate, depth of cut, and spindle speed) in this research. The algorithm is hybridized with Particle Swarm Optimization (PSO) to discover better solution in exploring the search space. Objective function of previous research is used to optimize the machining parameters in turning operation. The optimal machining cutting parameters estimated by FA that lead to a minimum surface roughness are validated using ANOVA test.

Full Text Available After an introduction on the basic aspects of electric railway transports, focusing mainly on driverless subways and their related automation systems (ATC, ATP, and ATO, a technique for energy optimization of the train movement through their control using genetic algorithms will be presented. Genetic algorithms are a heuristic search and iterative stochastic method used in computing to find exact or approximate solutions to optimization problems. This optimization process has been calculated and tested on a real subway line in Milan through the implementation of a dedicated Matlab code. The so-defined algorithm provides the optimization of the trains movement through a coast control table created by the use of a genetic algorithm that minimizes the energy consumption and the train scheduled time. The obtained results suggest that the method is promising in minimizing the energy consumption of the electric trains.

.... Genetic algorithms are then applied in conjunction with the finite element analysis to optimize the shape of cochlear implant electrode array based on the energy deposited in the spiral ganglion cells region...

The motivating application for this research is the global/local optimal design of composite aircraft structures such as wings and fuselages, but the theory and algorithms are more widely applicable...

To improve the tedious task of reconstructing gene networks through testing experimentally the possible interactions between genes, it becomes a trend to adopt the automated reverse engineering procedure instead. Some evolutionary algorithms have been suggested for deriving network parameters. However, to infer large networks by the evolutionary algorithm, it is necessary to address two important issues: premature convergence and high computational cost. To tackle the former problem and to enhance the performance of traditional evolutionary algorithms, it is advisable to use parallel model evolutionary algorithms. To overcome the latter and to speed up the computation, it is advocated to adopt the mechanism of cloud computing as a promising solution: most popular is the method of MapReduce programming model, a fault-tolerant framework to implement parallel algorithms for inferring large gene networks. This work presents a practical framework to infer large gene networks, by developing and parallelizing a hybrid GA-PSO optimization method. Our parallel method is extended to work with the Hadoop MapReduce programming model and is executed in different cloud computing environments. To evaluate the proposed approach, we use a well-known open-source software GeneNetWeaver to create several yeast S. cerevisiae sub-networks and use them to produce gene profiles. Experiments have been conducted and the results have been analyzed. They show that our parallel approach can be successfully used to infer networks with desired behaviors and the computation time can be largely reduced. Parallel population-based algorithms can effectively determine network parameters and they perform better than the widely-used sequential algorithms in gene network inference. These parallel algorithms can be distributed to the cloud computing environment to speed up the computation. By coupling the parallel model population-based optimization method and the parallel computational framework, high

Background To improve the tedious task of reconstructing gene networks through testing experimentally the possible interactions between genes, it becomes a trend to adopt the automated reverse engineering procedure instead. Some evolutionary algorithms have been suggested for deriving network parameters. However, to infer large networks by the evolutionary algorithm, it is necessary to address two important issues: premature convergence and high computational cost. To tackle the former problem and to enhance the performance of traditional evolutionary algorithms, it is advisable to use parallel model evolutionary algorithms. To overcome the latter and to speed up the computation, it is advocated to adopt the mechanism of cloud computing as a promising solution: most popular is the method of MapReduce programming model, a fault-tolerant framework to implement parallel algorithms for inferring large gene networks. Results This work presents a practical framework to infer large gene networks, by developing and parallelizing a hybrid GA-PSO optimization method. Our parallel method is extended to work with the Hadoop MapReduce programming model and is executed in different cloud computing environments. To evaluate the proposed approach, we use a well-known open-source software GeneNetWeaver to create several yeast S. cerevisiae sub-networks and use them to produce gene profiles. Experiments have been conducted and the results have been analyzed. They show that our parallel approach can be successfully used to infer networks with desired behaviors and the computation time can be largely reduced. Conclusions Parallel population-based algorithms can effectively determine network parameters and they perform better than the widely-used sequential algorithms in gene network inference. These parallel algorithms can be distributed to the cloud computing environment to speed up the computation. By coupling the parallel model population-based optimization method and the parallel

This paper presents a method for developing control laws for nonlinear systems based on an optimal control formulation. Due to the nonlinearities of the system, no analytical solution exists. The method proposed here uses the 'black box' structure of a neural network to model a feedback control law. The network is trained with the back-propagation learning method by using examples of optimal control produced with a differential dynamic programming technique. Two different optimal control prob...

Optimization is frequently employed in biomechanics research to solve system identification problems, predict human movement, or estimate muscle or other internal forces that cannot be measured directly. Unfortunately, biomechanical optimization problems often possess multiple local minima, making it difficult to find the best solution. Furthermore, convergence in gradient-based algorithms can be affected by scaling to account for design variables with different length scales or units. In this study we evaluate a recently-developed version of the particle swarm optimization (PSO) algorithm to address these problems. The algorithm’s global search capabilities were investigated using a suite of difficult analytical test problems, while its scale-independent nature was proven mathematically and verified using a biomechanical test problem. For comparison, all test problems were also solved with three off-the-shelf optimization algorithms—a global genetic algorithm (GA) and multistart gradient-based sequential quadratic programming (SQP) and quasi-Newton (BFGS) algorithms. For the analytical test problems, only the PSO algorithm was successful on the majority of the problems. When compared to previously published results for the same problems, PSO was more robust than a global simulated annealing algorithm but less robust than a different, more complex genetic algorithm. For the biomechanical test problem, only the PSO algorithm was insensitive to design variable scaling, with the GA algorithm being mildly sensitive and the SQP and BFGS algorithms being highly sensitive. The proposed PSO algorithm provides a new off-the-shelf global optimization option for difficult biomechanical problems, especially those utilizing design variables with different length scales or units. PMID:16060353

Full Text Available Community structure is one of the most fundamental and important topology characteristics of complex networks. The research on community structure has wide applications and is very important for analyzing the topology structure, understanding the functions, finding the hidden properties, and forecasting the time-varying of the networks. This paper analyzes some related algorithms and proposes a new algorithm—CN agglomerative algorithm based on graph theory and the local connectedness of network to find communities in network. We show this algorithm is distributed and polynomial; meanwhile the simulations show it is accurate and fine-grained. Furthermore, we modify this algorithm to get one modified CN algorithm and apply it to dynamic complex networks, and the simulations also verify that the modified CN algorithm has high accuracy too.

The authors present a class of efficient algorithms for PAC learning continuous functions and regressions that are approximated by feedforward networks. The algorithms are applicable to networks with unknown weights located only in the output layer and are obtained by utilizing the potential function methods of Aizerman et al. Conditions relating the sample sizes to the error bounds are derived using martingale-type inequalities. For concreteness, the discussion is presented in terms of neural networks, but the results are applicable to general feedforward networks, in particular to wavelet networks. The algorithms can be directly adapted to concept learning problems.

In this paper, the optimal reconfiguration of smart distribution system is performed with the aim of active power loss reduction and voltage stability improvement. The distribution network is considered equipped with wind turbines and solar cells as Renewable DG’s (RDG’s). Because of the presence of smart metering devices, the network state is known accurately at any moment. Based on the network conditions (the amount of load and generation of RDG’s), the optimal configuration of the network is obtained. The optimization problem is solved using a recently introduced method known as Grey Wolf Optimizer (GWO). The proposed approach is applied on 69-bus radial test system and the results of the GWO are compared to those of Particle Swarm Optimization (PSO), and Genetic Algorithm (GA). The results show the effectiveness of the proposed approach and the selected optimization method.

Currently information on the web is accessed through different devices. Each device has its own properties such as resolution, size, and capabilities to display information in different format and so on. This calls for adaptation of information presentation for such platforms. This paper proposes...... content-optimized and time-optimizedalgorithms for information presentation adaptation for different devices based on its hierarchical model. The model is formalized in order to experiment with different algorithms....

We explore the applicability of the differential evolution algorithm in finding the global minima of three typical nuclear structure physics problems: the global deformation minimum in the nuclear potential energy surface, the optimization of mass model parameters and the lowest eigenvalue of a nuclear Hamiltonian. The algorithm works very effectively and efficiently in identifying the minima in all problems we have tested. We also show that the algorithm can be parallelized in a straightforward way.

Full Text Available A new parallel variable distribution algorithm based on interior point SSLE algorithm is proposed for solving inequality constrained optimization problems under the condition that the constraints are block-separable by the technology of sequential system of linear equation. Each iteration of this algorithm only needs to solve three systems of linear equations with the same coefficient matrix to obtain the descent direction. Furthermore, under certain conditions, the global convergence is achieved.

In permanent magnet wigglers, magnetic imperfection of each magnet piece causes field error. This field error can be reduced or compensated by sorting magnet pieces in proper order. We showed a genetic algorithm has good property for this sorting scheme. In this paper, this optimization scheme is applied to the case of permanent magnets which have errors in the direction of field. The result shows the genetic algorithm is superior to other algorithms.

Non-traditional search and optimization methods based on natural phenomena have been proposed recently in order to avoid local or unstable behavior when run towards an optimum state. This paper describes the principles of bio inspired algorithms and reports on Migration Algorithms and Bees...

of the cache misses by a factor of d compared to the unidirectional algorithm which is the common standard up to now. The new algorithm is also optimal in the sense that the leading term of the cache misses is reduced to scanning complexity, i.e., every degree of freedom has to be touched once. We also present...

In this paper we present a generic optimizationalgorithm for the allocation of dynamic positioning actuators, such as azimuthing thrusters and fixed thrusters. The algorithm is based on the well-known Lagrange multipliers method. In the present approach the Lagrangian functional represents not only