Navneuronet.narod.ru

Kalman extension of the genetic algorithm
Los Alamos National Laboratory, MS-F607, Los Alamos, NM 87545 LAUR-00-150, SPIE proc. vol. 4055, April, 2000 ABSTRACT
In typical GA application, the fitness assigned to a chromosome-represented individual in the context of a specified environment takes a deterministic calculable value. In many problems of interest, the fitness of an individual is stochastic, and the environment changes in unpredictable ways. These two factors contribute to an uncertainty that can be associated with the estimated fitness of the individual.
The Kalman formulation provides a mechanism for a useful treatment of uncertainty within the GA framework. It provides for tracking the best-estimated fitness, and for assessing the uncertainty of the best-estimated fitness. The stochastic uncertainty of an existing individual can be reduced by additional evaluation of the fitness. The process (environmental) noise causes an increase in uncertainty as the age of the last evaluation increases. In a Kalman-extended genetic algorithm, we want to know the value of re-evaluating an existing solution in the population relative to generating and evaluating new individuals. A scheme for efficient allocation of computational resources among existing individuals and new individuals is This Kalman-Darwin formulation is applied to the problem of maintaining an optimal network configuration of links between moving nodes, with time-dependent, stochastic blockages. The nodes, for example, could be environmental sensors with radio transmitter/relays located on vehicles, and the network is configured to provide communication paths from each sensor back to a central node with minimum message loss. As the sensors move, the optimal network changes, but information contained within the GA population of solutions allows new optima to be efficiently obtained. The performance of this approach is explored and results are presented.
1. INTRODUCTION
The Kalman Genetic Algorithm (KGA) is an extension of the genetic algorithm (GA)1-3 in which Kalman filtering4-6 is applied to the fitness values associated with the population of chromosomes. In the GA, a chromosome represents a trial solution to a problem. The chromosome has a well-defined measure of fitness within the context of the problem space or environment. In many interesting applications, the environment is dynamic. The fitness of a given chromosome will change in time, as changes occur in the environment. In addition, for environments of sufficient complexity, or for environments with stochastic elements, the fitness of a chromosome is unknowable, and can only be approximated. The Kalman formulation provides a natural mechanism for extending the domain of application of the genetic algorithm from static, prescribed problem spaces to dynamic, and complex or stochastic environments.
In the KGA, at any given time, each chromosome has a best-estimated fitness and an uncertainty associated with that estimate. The fitness and uncertainty are updated according to the Kalman formulation, as the environment changes with the * Correspondence: E-mail: [email protected]; Telephone: 505-667-6654 passage of time, and when a chromosome is re-evaluated. Knowledge is acquired7 by re-evaluating an existing trial solution (thereby decreasing the uncertainty associated with its fitness estimate) or by generating and evaluating a new trial solution.
The principle behind the KGA is simple. The population, consisting of N individuals, each with a best-estimated fitness, can be characterized by the variance V of these N fitness values. When the uncertainty associated with all members of the population is smaller than V, the population is treated as sufficiently evaluated, and the algorithm will generate and evaluate a new trial solution. However, if any of individuals in the population have uncertainties greater than V, the algorithm will re- evaluate whichever individual has the highest uncertainty.
The literature contains applications in which GA is used to tune the parameters of a Kalman filter.8-11 There is also some research reported as a hybridization of Kalman filtering and GA, in which Kalman filtering and GA are used to optimize two separate parts of a problem.12,13 In previous work,14-17 the author used GA to periodically optimize controllers in the context of a dynamic environment. The heuristic of including the previous best solution in the next initial population was explored, but a theoretical formulation underlying the use of GA in noisy, dynamic environments was lacking. In a SciSearch® search of 17.5 million technical articles, no reference was found in which a Kalman filter is used as an integral part of the GA 2. MECHANICS OF THE KALMAN GENETIC ALGORITHM
The KGA keeps a population of individuals. An individual represents a trial solution to a problem. There is an upper limit, N, on the population size, and at any given time, there are actually n individuals in the population. The individual, i, has four elements: 1) a chromosome W , 2) a best estimated fitness, f , 3) an uncertainty, P , associated with that estimate, and 4) a timestamp, T , specifying the time that the uncertainty was last updated. The chromosomes can take a variety of forms.18 The requirements on the form of the chromosome are that it can be transformed into a possible solution of the problem, and that a set of genetic operators can be constructed for it. For this implementation, the chromosomes are simply fixed length sequences of bounded integers. A mechanism for transformation from the chromosomal form (the genotype) to the problem solution form (the phenotype) must be developed for each problem. As an aid to interpretation of the uncertainty, the underlying fitness can thought of as the best estimate of fitness, plus or minus the square root of the uncertainty.
When a real-time directed-search algorithm is used in a dynamic environment, the algorithm computational time is interwoven with the physical time. We consider the case in which an evaluation of the fitness of a trial solution takes a finite amount of time, τ . During a given period of time, there will thus be a finite number of trial evaluations that the algorithm A mechanism is required to evaluate the fitness of a solution. In environments with stochastic processes, the result of a fitness evaluation will be a real-number value that represents the true, underlying fitness value but has an uncertainty associated with it. If an ensemble of fitness evaluations were to be made on a trial solution, the variance of the ensemble of the resulting fitness values, R, is taken as the measure of the uncertainty in the fitness evaluation. This variance is analogous to the observation noise variance of the Kalman formulation, if a fitness evaluation is treated like an observation. R can be interpreted as the uncertainty with which the result of a fitness evaluation represents the true, underlying fitness of the trial solution. R designates the variance of the observation noise associated with a single evaluation of a trial solution.
Because the environment is dynamic, the underlying fitness of a given trial solution will change with time. In the Kalman formulation, this change is attributed to process noise. Like the observation noise, the process noise is characterized by a variance, in this case designated by Q. Q gives the variance in the change in the underlying fitness of trial solutions during a specified time interval, due to change in the environment. In this formulation, there is no fixed time interval to use to define Q, as is common in Kalman filtering applications. Instead, a “process noise rate”, given by q, is used. The process noise variance expected after the environment undergoes changes for a time interval dt, is Q = dt q. This presumes that the process noise variance increases linearly with time interval, although alternative formulations could be used as well. As with any Kalman application, q can be specified by knowledge about the environment, or it can be evaluated adaptively.
The KGA begins operating at time t = t . At this time, the population contains no trials yet (n = 0). The algorithm iterates over four main processes, as shown schematically in Fig. 1. The first process updates the uncertainties of the members of the population to the current time. If an uncertainty was last updated at time T , the associated uncertainty increases according to This is analogous to the Kalman equation that updates the covariance matrix for process noise. The time of the last update is then advanced to the current time, T = t.
Figure 1. The Kalman-extended genetic algorithm, shown schematically as a cycle over four major processes.
The second process characterizes some statistics of the population that will determine whether to generate a new trial, or re- evaluate an existing individual. The variance, V, of the best-estimated fitnesses of the individuals in the population is , of the most uncertain individual. These two statistics are used by the algorithm to determine which of the three elements of the third process to invoke.
The third process performs a fitness evaluation. There are three alternatives for this third process. If the population is not full (i.e. n<N), a new random chromosome is generated and evaluated. The second alternative for the evaluation process: if the population is full, and the uncertainty of the most uncertain individual exceeds the population’s fitness variance (i.e. P the most uncertain individual will be reevaluated. When a re-evaluation is performed (returning a value of g), the previous , is updated using the Kalman observation equation: In addition, a re-evaluation reduces the uncertainty P by multiplying by a factor of R/(P+R) as follows: The quantity P/(P+R) is analogous to the gain, K, of the Kalman formulation.
The third alternative for the evaluation process: if the population is full, and the uncertainty of the most uncertain individual is less than the population’s fitness variance, a new trial will be generated and evaluated. Traditional GA mechanics are used to generate new individuals. Two parents are selected based on their fitness, with more fit individuals being more likely to be selected. Genetic operators such as crossover and mutation are used to generate a new chromosome from those of the parents.
When a new chromosome is created and evaluated once, the best estimate of fitness f is the result of the evaluation, and the The fourth process incorporates the newly evaluated trial appropriately into the population. The best-estimated fitness and the uncertainty are combined to give an uncertainty-compensated fitness, F, for each individual, given by The algorithm thus searches for individuals with high fitness, but discounts the fitness if there is a high associated uncertainty. The population is sorted by this uncertainty-compensated fitness. If the uncertainty-compensated fitness of the new chromosome is better than that of the least fit individual in the population, the new individual is added to the population, and the least fit individual is discarded. If the population is not yet full, or if an existing individual was reevaluated, the population is simply re-sorted according to uncertainty-compensated fitness. At any time, the individual with the best uncertainty-compensated fitness is taken to be the best-adapted solution produced by the algorithm. These four processes take some time to perform, but the bulk of the time is presumably spent evaluating trials. When the cycle repeats, the time t is 3. DEVELOPMENTAL TEST-BED: NETWORK CONFIGURATION OPTIMIZATION
A test-bed has been constructed for developing and experimenting with the KGA in a simulated dynamic, noisy environment.
It has been implemented in Java, with a GUI that allows a user to easily construct scenarios and observe simulations of the The test problem is to configure a network of radio links connecting a set of nodes to a collection point. The task is to maintain a good configuration of links as the nodes move. This test application derives from a system of atmospheric aerosol sensors mounted on vehicles, which deliver their data via radio links to a collection point. Each sensor collects a stream of data, which it then transmits via a low-power radio link. A sensor node may transmit directly to the collection point or to another node. A node can receive and repeat data from other sensors. The network of links thus is an acyclic, directed graph (i.e. a tree) with the base being the collection point, and the sensors forming the other nodes. Each radio link forms an edge of In its pure, unconstrained form, this problem admits an algorithmic solution (via Dijkstra’s shortest-path tree algorithm).19-21 The existence of this solution does not render the application trivial, because slight variations in the problem definition, as would be encountered in practice, make this problem NP-complete.19 Examples of such variations include: a constraint on the number of links a given node may receive and retransmit; a bandwidth limitation on the transmission channels or a degradation in transmission fraction with increased message traffic; a maximum number of links in a path to the collection point; or the possibility of redundant paths, where a node may transmit to more than one receiver. Optimization of a complex network with routing and bandwidth constraints is not amenable to algorithmic approaches, but can be attacked with GA.22 The existence of an algorithmic solution for the pure form of the problem does allow the performance of the KGA to be assessed relative to the sequence of optimal solutions.
The nodes are treated as executing a random walk within a prescribed area. (The vehicles are not dedicated to the sensors, but are engaged in other activities.) The movement of the nodes makes the problem environment dynamic. As the nodes move, the transmission probability along the links changes, and the optimal network configuration also changes.
A prescribed area on the ground of arbitrary shape and size is represented on a 2D Cartesian cell grid. The cell size, and the number of cells spanning the two dimensions, are adjustable. A baseline scenario has been constructed, in which each cell represents a 160 by 160 meter square, and a grid of 125 by 125 cells covers a 20 by 20 km area. In this baseline scenario, the vehicle/sensor/nodes are restricted to a roughly 10 by 20 km elliptical region as shown in Fig 2.
Figure 2. Example scenario, showing a user-prescribed region, 25 randomly placed nodes, and a randomly generated link network. The collection point is shown as the open circle.
Let S designate the total number of nodes, exclusive of the collection point. The baseline scenario contains 25 sensor nodes.
These vehicle-mounted sensors can be initialized by assigning each sensor to a random cell within the prescribed area. Only one sensor can occupy a cell. Fig. 2 shows the locations of 25 randomly placed sensors.
Each node establishes a radio link to another node or to the collection point, to transmit its own data, and any data it receives from other nodes. Each node has a unique ID in the range of one to S. The receiver of a node is specified by an integer in the range of zero to S. A value of zero indicates that the link is transmitting directly to the collection point, while any other value specifies the ID of the receiving node. The network configuration is completely specified by giving the receiver for each node.19 This requires a total of S integers, each in the range 0 to S. A node can not have itself as a receiver. The number of distinct network configurations is then SS. For 25 sensors, there are 8.9(10)34 possible network configurations.
In this implementation, only valid spanning-trees configurations (including the collection point) are considered, i.e. all nodes and the collection point are included in the graph, and there are no cyclic paths in the graph. This reduces the size of the problem space. A simple algorithm is used to generate random trees, as follows. All nodes are initially unconnected. An unconnected node is selected at random. A receiver is selected at random from the set containing the collection point and all connected nodes. The selected unconnected node takes the randomly selected receiver to be its receiver, and then becomes a connected node. This process is repeated until all nodes are connected. A randomly generated network linking 25 sensors to a collection point is shown in Fig. 2. The 25 integer representation of this network is { 8 19 17 18 12 20 20 20 11 0 21 0 20 20 13 12 12 20 24 12 0 6 18 12 21}. Node 1 transmits to node 8, while nodes 10, 12 and 21 transmit to the collection point, etc.
The expected transmission fraction over a link depends on the length of the link. For short distances, the expected transmission is nearly perfect, because the signal is much higher than the noise. There is a scale distance, d, where the expected transmission fraction across a link is reduced to one half; for significantly longer transmission distances, the transmission fraction will be very small (extinction and the 1/r2 antenna gain relation degrades the signal, which becomes obscured in noise.) There is another scale distance, w, which characterizes the distance over which the transmission fraction drops from near one to near zero. A parameterized formulation is used for the transmission fraction from node i to node j: where d designates the distance from node i to node j. The transmission as a function of distance is shown in Fig. 3 for d = 5000 m and w = 100 m. If the path from a node to the collection point contains a sequence of links, the total transmission fraction from the node to the collection point is simply the product of the transmission fractions of each link in the path.
Fig. 3. The transmission fraction across a link, as a function of the length of the link, using the parameterized formulation with d=5000 m, The fitness of a network configuration can be characterized in a variety of ways, depending on what is important in any given application. For the test-bed application, the network configuration fitness is simply the average, over all nodes, of the transmission loss from each node to the collection point. The goal of maximizing transmission fraction is then equivalent to a goal of minimizing message transmission loss. For the random network shown in Fig. 2, the average message loss is 0.94594.
The shortest-path algorithm was used to generate the network that minimizes the transmission losses from each node to the collection point. The shortest-path tree algorithm minimizes the sum of the weights associated with all edges in the paths from any node to the root node. The algorithm can be used by equating the weight associated with an edge to the natural log of the inverse of the transmission fraction. Then minimizing the sum of the weights along a path is equivalent to maximizing the total transmission fraction along the path. The shortest-path network is shown in Fig 4, where the sensors have the same locations as in Fig. 2. For this network configuration, the average message loss fraction is 0.03185, which is representative of 25 sensors in a 10 by 20 km elliptical area, where the transmission range scale is 5 km.
Fig. 4. The shortest-path tree for 25 randomly located nodes.
The dynamic environment is implemented as follows. A simulation is run in which the time is advanced in steps of durationτ . During each time step, the fitness of one network configuration is evaluated, as prescribed by the KGA. In addition, during each time step, a sensor may move, causing a change in the environment. In order to explore sensitivity of the KGA to process noise, the expected change in fitness during a time step can be varied by controlling how often sensors are moved.
The process noise rate is then calculated adaptively, using a fixed-gain recursive filter and the observed fitness changes caused by the sensor movements. If, for example, sensors are moved (to one of the eight nearest-neighbor cell locations) at a rate of one movement per 500 time steps, the process noise variance is found to average approximately 8.51E-11 in a time step. During the 100 time steps which would be required to re-evaluate all members of a 100-member population, the process noise would be ~10E-8, corresponding to a typical fitness change of ~0.0001. This process noise level is about one part in 300 of typical near-optimal network configurations in the baseline scenario. In a simulation starting with the sensor configuration shown in fig. 2 and incurring 500 sensor movements, the fitness of the shortest-path tree network configuration varies over the range of 0.0295 to 0.0335. If the links in the initial shortest-path network configuration were left unchanged during these 500 sensor movements, the fitness would be significantly degraded.
The observation noise, R, is implemented as follows. The transmission fraction over a link depends not only on the length of the link, but also on stochastic factors such as the atmospheric conditions and line-of-sight obstructions. These stochastic factors contribute to the observation noise that gives variation in the actual fitness of a network configuration. In the test-bed, process noise is added to average transmission fraction of a network configuration whenever it is evaluated. An evaluation of the fitness of a given network configuration for a given set of node positions is artificially given an error, by adding (2r-1)sqrt(3R), where r is a random number between 0 and 1. This produces a distribution of observation noise with a An implementation of the genetic algorithm is used which takes a sequence of integers as its chromosome, where the integers are bounded in zero to a specified maximum value. In this case, the chromosome has S integers, each in the range of [0, S].
The population size is 100 individuals. The first 100 individuals are created as random tree configurations. When the algorithm determines that a new trial solution should be generated, two parents are selected. This selection is based on fitness rank, with the most fit individual being 2.5 times more likely to be selected than the median fit individual. Generation of new individuals uses two-point cross-over, followed by mutation and a repair step. The cross-over causes complete genes (i.e.
integers in 0 to S) to be taken from the parents into the child: there is no splitting of genes. The mutation step selects two genes at random, and resets them to new random values, corresponding to a receiver node selected at random from those connected to the collection point. The repair step ensures that the resulting network configuration is an acyclic spanning tree rooted on the collection point. Every gene that is part of a valid tree is retained – those that are not are replaced with alleles that correspond to a link to a connected node. If the algorithm determines that an existing individual should be re-evaluated, 4. RESULTS
A simulation was run in which the observation noise variance, R, of the fitness (the message loss fraction) was set to 1E-8.
Since the message loss fraction of good configurations is around 0.03, the standard deviation of the observation noise of 1E-4 is about one part in 300, which can be considered to be a low level of observation noise. For this case, the process noise variance is approximately 8.51E-11 per time step, which is achieved by moving the sensors at the rate of one movement per 500 time steps. As described above, for a population size of 100, this process noise standard deviation is below one part in 300, which can be considered to be a low level of process noise. Fig 5 shows the fractional error of the actual fitness (average message loss rate, with observation noise removed) of the best member (lowest uncertainty-compensated message loss fraction) of the population relative to the fitness of the shortest-path network configuration, i.e. (f-f )/f . Initially, when the KGA has evaluated only one random configuration (that shown in Fig. 2) the loss of the best configuration is 0.94594, while the loss with the optimal configuration is 0.03185, so the fractional error of the KGA relative to optimal has a high initial value of 2870%. As the KGA generates and evolves its population of trial solutions, the fractional error is seen to drop. After about 20,000 trial evaluations, the best solution in the KGA population is within 0.1% of the optimal solution. In this case, the KGA often attains the optimal solution.
During the 250,000 time steps of the simulated run, the KGA evaluated 100 random configurations, generated and evaluated 55,674 new configurations, and re-evaluated existing configurations 194,226 times. When the observation noise is low, the algorithm spends 22% of its resources on new trial solutions, and 78% on re-evaluating existing solutions. Note that once a population has converged somewhat on the optimal configuration, when a new trial is generated, it will have a larger uncertainty than the bulk of the population. If the fitness of the new trial is good enough to earn it a place in the population, the algorithm will then re-evaluate the new trial until its uncertainty drops below the population variance.
Several events can be seen in fig. 5 as spikes in the fractional error. These occur when the motion of the sensors causes the optimal network configuration to undergo a significant restructuring. The fractional error between the best KGA solution and the optimal solution may rise to a few percent on these occasions, but then drops back down as the KGA discovers the new configuration. Note that these recoveries require one or two thousand time steps, compared to the 20,000 time steps required to reach near-optimal solutions beginning from scratch.
Number of Evaluations
Fig. 5. The fractional error of the message loss fraction of the best KGA solution relative to that of the optimal network configuration, showing how closely the KGA tracks the optimal network configuration. There is one trial evaluation per time step. The process noise variance is approximately 10E-10 per timestep, and observation noise variance is 10E-8, for 25 sensor network starting from the locations shown in Fig. 2. The loss fraction of the optimal configuration ranges from 0.0295 to 0.0334.
A second case was run in which the process noise was increased by a factor of about 100. This was accomplished by moving the sensors at a rate of one sensor motion every 50 time steps, which results in a process noise variance that averaged 9.26E-9 per time step. This corresponds to a process noise standard deviation of about 0.001 in 100 time steps (one for each member of the population), which is about 3% of the fitness of near-optimal configurations. As shown in fig. 6, this increased process noise degrades the ability of the KGA to track the optimal network configuration. Where the optimal configuration has a message transmission fraction that varies from 0.0264 to 0.0357, with an average value of 0.02978 (excluding the first 20,000 time steps), the KGA has an average message loss fraction of 0.03093 during the same period. On average the KGA solution is 3.9% worse than the optimal configuration, excluding the first 20,000 time steps, at this process niose level of 3%. For this case, there is little change in the resource allocation between re-evaluation and new trial generation: the KGA called for new trial solutions in 21.1% of time steps.
A third case was run which was the same as the first, except that the observation noise variance was increased by a factor of 100, to a moderate level of 1.E-6. The process noise remained at the low level of 1E-10 per time step. The observation niose standard deviation (0.001) is about 3% of the fitness of near-optimal configurations for this third case. Fig. 7 shows how the KGA is able to track the optimal configuration at this noise level. After an initial period of 20,000 time steps, the KGA configuration is on average 1.92% worse than the optimal configuration. At this observation noise level, significantly more resources are allocated to re-evaluation of existing trials: new trial configurations are generated and evaluated in only 8.1% of Number of Evaluations
Fig. 6. The fractional error of the message loss fraction of the best KGA solution relative to that of the optimal network configuration. The case is identical to that shown in Fig 5, except that the process noise variance is increased by a factor of 100 to a value of approximately Number of Evaluations
Fig. 7. The fractional error of the message loss fraction of the best KGA solution relative to that of the optimal network configuration. The case is identical to that shown in Fig 5, except that the observation noise variance, R, is increased by a factor of 100 to a value of 5. CONCLUSIONS
The overall conclusion of this work is that the Kalman formulation can be integrated into the genetic algorithm formulation in a compelling way to create a Kalman-extended genetic algorithm that can perform ongoing directed searches in dynamic, noisy environments. The extension of a population of individuals, each with a fitness, by associating an uncertainty with the fitness of each individual, is conceptually clear, and easy to implement. The mechanics of updating the best-estimated fitness and the uncertainty after a re-evaluation follow directly from the Kalman formulation. The very simple implementation of the KGA, where a new trial is generated and evaluated whenever the uncertainty of the most uncertain individual is less than the variance of the population fitnesses, has been found to work, through experimentation in a test-bed.
The KGA can be viewed as a mechanism to allocate resources between the two complimentary activities of re-evaluating existing solutions for changing conditions, and generating new solutions. The allocation depends on the level of observation noise: the more uncertain the observations are, the more resources must be dedicated to re-evaluation of existing solutions.
Even when observation noise is low, however, a somewhat surprising result is that only 21 or 22% of effort should be expended on new solutions, while the remainder should go to re-evaluating existing solutions.
At low levels of process and observation noise (±0.3%), the KGA was shown to be able to maintain near-optimal solutions.
At moderate observation noise level (±3%), the KGA was still able to maintain configurations with fitness values within an average of 2% of the optimal configuration. At moderate process noise level (±3%), the KGA was able to maintain configurations with fitness values within an average of 4% of the optimal configuration.
In contrast to the traditional GA wherein a bigger population is better, in the KGA there is an optimal population size. If the population were too large, the KGA would waste time evaluating relatively poor individuals that have grown obsolete (i.e.
whose uncertainty has increased too much). There are indications that the optimal population size can be re-evaluated on an ongoing basis, using R, q τ , and V.
The formulation presented here was developed with simplicity as a criterion. Alternative forms may be more effective. For example, the uncertainty-compensated fitness, which is used to rank the individuals in the population, could be a more complex function of the best-estimated fitness, the uncertainty, and/or the square root of the uncertainty. The criteria for selecting which individual to evaluate next could depend on the best-estimated fitness in addition to the uncertainty. The population fitness variance could also be formulated differently.
When a new trial is generated and evaluated, its uncertainty is generally much higher than that of the existing individuals.
The new trial should therefore be re-evaluated several times to reduce its uncertainty to a level at which a valid comparison can be made to the existing individuals. Less fit trials can be discarded at higher level of uncertainty, improving the efficiency of the algorithm. This is an avenue for improvement of the formulation presented here, in which a new trial is compared with the population after only a single fitness evaluation.
REFERENCES
J. H. Holland, Adaptation in Natural and Artificial Systems, MIT Press, Cambridge, MA (1992).
D. E. Goldberg, Genetic Algorithms in Search, Optimization and Machine Learning, Addison-Wesley Publishing Co.
David B. Fogel, Evolutionary Computation : Toward a New Philosophy of Machine Intelligence, IEEE Press (1995).
R. E. Kalman, “A new approach to linear filtering and prediction problems,” Trans. AMSA Journal of Basic Engineering, vol. 82, pp 35-45 (1960).
R. E. Kalman and R. S Bucy, “New results in linear filtering and prediction theory,” Trans. AMSA Journal of Basic Engineering, vol. 82, pp 95-108 (1961).
S. Haykin, “Adaptive filter theory,” Prentice-Hall Information and System Science Series (1986).
H. Plotkin, Darwin Machines and the Nature of Knowledge, Harvard University Press, Cambridge, MA (1994).
P. E. Howland, “Target tracking using television-based bistatic radar,” IEE Proc. Radar Sonar and Navigation, v.
M. J. Arcos, C. Alanso, and M. C. Ortiz, “Genetic-Algorithm-Based Potential Selection in Multivariant Voltammetric Determination of Indomethacin and Acemethacin by Partial Least-squares, “ Electrochimica Acta, v. 43(#5-6) pp.
W. S. Chaer, R. H. Bishop, and J. Ghosh, “A Mixture-of-Experts Framework for Adaptive Kalman Filtering,” IEEE trans. On Systems, Man, and Cybernetics, Part B-Cybernetics , vol. 27, no. 3, pp. 452-464 (June 1997).
K. G. Berketis, S. K. Katsikas, and S. D. Likothanassis, “Multimodal Partitioning Filters and Genetic Algorithms,” Nonlinear Analysis-Theory Methods & Applications, vol. 30, no. 4, pp. 2421-2427 (Dec 1997).
L. A. Wang, J. Yen, “Extracting Fuzzy Rules for System Modeling using a Hybrid of Genetic Algorithms and Kalman Filter,” Fuzzy Sets and Systems, vol. 101, no. 3, pp. 353-362 (Feb 1, 1999).
S. E. Aumeier, J. H. Forsmann, “Evaluation of Kalman Filters and Genetic Algorithms for Delayed-Neutron Nondestructive Assay Data Analyses,” Nuclear Technology, v. 122, no. 1, pp. 104-124 (April 1998).
Phillip D. Stroud, “Evolution of Cooperative Behavior in Simulation Agents,” SPIE proc. vol. 3390 (April 1998).
Phillip D. Stroud, “Adaptive Simulated Pilot,” Journal of Guidance, Control, and Dynamics, Eng. notes, vol. 21, no. 2, Phillip D. Stroud, “Learning and Adaptation in an Airborne Laser Fire Controller,” IEEE Transactions on Neural Networks, vol.8, no. 5, pp. 1078-1089 (Sept 1997).
Phillip D. Stroud, and Ray C. Gordon, “Automated Military Unit Identification in Battlefield Simulation,” SPIE proc.
Z. Michalewicz, “Genetic Algorithms + Data Structures = Evolution Programs,” 2nd ed., Springer, Berlin (1992).
L. J. Dowell, “Optimal Configuration of a Command and Control Network: Balancing Performance and Reconfiguration Constraints,” Proceedings of the 2000 ACM Symposium on Applied Computing (March 2000).
E. W. Dijkstra, “A Note on Two Problems in Connexion with Graphs,” Numerische Mathematik 1, pp. 269-271 E. Minieka, “Optimization algorithms for networks and graphs,” M. Dekker, New York (1978).
King-Tim Ko, Kit-Sang Tang, Cheung-Yau Chan, Kim-Fung Man, Sam Kwong, “Using Genetic Algorithms to Design Mesh Network,” Computer, Vol. 30, No. 8 (August 1997).

Source: http://navneuronet.narod.ru/ftp/Los_Alamos/Phillip_Stroud/KGA00150.pdf

Untitled document

PRODUCT LIST TABLE OF CONTENTS Antibiotics and Related Products Analgesic/ Antipyretics/ Antinflammatory Anti-Asthmatics/ Antihistamines Anti-Diarrhoea & Laxatives Anti-Rheumatics/ Steroids Cough & Cold Preparations External Preparations Anti-Fungal Gastrointestinal Preparations Vitamins & Minerals Anti-Diabetics Anti-Hypertensive Miscellan

Microsoft word - smithers_latestnn.doc

Smithers Summary Shandi Hopkins was admitted to the residential treatment program, Smithers Alcoholism and Rehabilitation Center, at Roosevelt Hospital on 4/18/01 and discharged 5/27/01. He was in touch with me regularly, sometimes twice a day while in the program. I obtained an 800 number to enable him to stay in touch with me when I moved from New York last July to begin teaching at Illinoi

Copyright © 2012-2014 Medical Theses