Discovery and Optimization of Materials Using Evolutionary

May 12, 2016 - ... akin to Darwinian evolution: survival of the fittest. Somewhat surprisingly, the materials science research community has not made ...
0 downloads 0 Views 23MB Size
Review pubs.acs.org/CR

Discovery and Optimization of Materials Using Evolutionary Approaches Tu C. Le† and David A. Winkler*,†,‡,§,∥ †

CSIRO Manufacturing, Bag 10, Clayton South MDC, Victoria 3169, Australia Monash Institute of Pharmaceutical Sciences, 381 Royal Parade, Parkville 3052, Australia § Latrobe Institute for Molecular Science, La Trobe University, Bundoora 3046, Australia ∥ School of Chemical and Physical Sciences, Flinders University, Bedford Park 5042, Australia ‡

ABSTRACT: Materials science is undergoing a revolution, generating valuable new materials such as flexible solar panels, biomaterials and printable tissues, new catalysts, polymers, and porous materials with unprecedented properties. However, the number of potentially accessible materials is immense. Artificial evolutionary methods such as genetic algorithms, which explore large, complex search spaces very efficiently, can be applied to the identification and optimization of novel materials more rapidly than by physical experiments alone. Machine learning models can augment experimental measurements of materials fitness to accelerate identification of useful and novel materials in vast materials composition or property spaces. This review discusses the problems of large materials spaces, the types of evolutionary algorithms employed to identify or optimize materials, and how materials can be represented mathematically as genomes, describes fitness landscapes and mutation operators commonly employed in materials evolution, and provides a comprehensive summary of published research on the use of evolutionary methods to generate new catalysts, phosphors, and a range of other materials. The review identifies the potential for evolutionary methods to revolutionize a wide range of manufacturing, medical, and materials based industries.

CONTENTS 1. Introduction 2. Evolutionary Algorithms 2.1. Genetic Algorithms 2.1.1. Overview of the Process 2.1.2. Materials Genomes 2.1.3. Fitness Functions 2.1.4. Fitness Landscapes 2.1.5. Genetic Operators 2.2. Structure−Property Model Methods as in Silico Fitness Functions 2.3. Genetic Algorithm Software and Optimization Parameter Choices 3. Examples of Materials Discovery and Optimization Using Evolutionary Algorithms 3.1. Catalytic Materials 3.1.1. Catalyst Evolution Using Experimental Fitness Functions 3.1.2. Catalyst Evolution Using Computational Models as Fitness Functions 3.2. Phosphors 3.3. Other Materials 4. Optimization Using Structure−Property Models of Materials Evolutionary Landscapes 5. Conclusions and Perspective Author Information

Corresponding Author Notes Biographies References

6107 6108 6109 6109 6109 6109 6110 6110

1. INTRODUCTION The vastness of materials structure space is still not widely recognized in the materials science community. Estimating how many new materials could be made from the elements in the periodic table using the laws of chemical valence and reactivity is very difficult. However, most studies have agreed on a number of close to 10100, considerably larger than the estimated number of particles of matter in the Universe. For all intents and purposes, the number of possible materials we could synthesize is infinite. This realization immediately raises a threat and an opportunity. The threat revolves around the impossibility of exhaustively exploring such vast space using even the most optimistic projections of the capabilities of robotics and automation. Although these high throughput materials synthesis and characterization capabilities are being developed at a rapid pace,1,2 and they will undoubtedly improve the efficiency of discovery of new and useful materials, they do not provide an answer to the materials space problem. The

6110 6112 6113 6113 6113 6118 6121 6123 6126 6129 6130

Published 2016 by the American Chemical Society

6130 6130 6130 6130

Received: November 26, 2015 Published: May 12, 2016 6107

DOI: 10.1021/acs.chemrev.5b00691 Chem. Rev. 2016, 116, 6107−6132

Chemical Reviews

Review

and effectively by performing a reasonable number of experiments. We trust that this comprehensive and timely review will be useful to a broad materials science audience and will inspire scientists to use these very efficient, state-of-the-art computational tools to make major discoveries in materials design and development. This paper is organized as follows. Section 2 gives a brief introduction to evolutionary methodologies. This is followed by a comprehensive review of studies on materials optimization using genetic algorithms, the most widely used technique, in section 2.1. The use of models of materials structure property (genotype−phenotype) relationships as surrogate fitness functions is described in section 2.2. Section 3 reviews the literature on the application of evolutionary algorithms to identify and optimize new materials. Sections 3.1, 3.2, and 3.3 are devoted to catalysts, phosphors, and other materials, respectively. Section 3.1 provides a thorough review of catalyst discovery and optimization using genetic algorithms, the largest body of published work. It also describes the use of both experimental and computational fitness functions to evolve materials into promising areas of catalyst space. Catalyst optimization studies that use genetic algorithms and, for example, models of catalyst structure−property relationships as surrogate fitness functions, offer benefits in time and cost savings compared to carrying out experiments alone and are reviewed in sections 3.2 and 3.3. Materials optimization using quantitative structure−property relationships models trained on data derived during evolutionary optimization is reviewed in section 4. Conclusions and prospects for the future of evolutionary methods in materials discovery and optimization are given in section 5.

opportunity that vast materials spaces offer is that, even if this space is sparsely populated by regions with useful properties, there will be an almost inexhaustible number of them, providing rich areas for research and application in the future. Exploration of materials spaces by computational rather than experimental means can be much faster and more efficient, and a growing body of published work supports the utility of computational methods in materials design and optimization.3−5 The key question is, how do we explore vast materials structure and property spaces as efficiently as possible using very fast computational methods? By drawing on resources from other areas of research, namely evolutionary mathematical methods, we have a rational and viable means of tackling this extremely challenging but also very important problem. These methods are some of the most effective ways of searching extremely large search spaces consisting of multiple dimensions.6−9 Evolutionary methods, where a population of promising materials is altered (mutated) to increase some useful property (fitness), are akin to Darwinian evolution: survival of the fittest. Somewhat surprisingly, the materials science research community has not made significant use of these very promising methods, possibly because much of the work is published in engineering and computer science journals, but also because the experimental methods to implement them (high throughput synthesis and characterization technologies) have only recently been developed and applied. Innovations in combinatorial chemistry in the pharmaceutical industries, high throughput peptide and oligonucleotide synthesis methods, and robotics and automation in general10−12 have stimulated recent efforts to automate materials synthesis and characterization. These are also important enabling technologies for wider and more effective utilization of evolutionary methods. Genetic algorithms have been used to solve optimization problems in some computational tools used in drug design, for example. Since the early 1990s there have been successful applications of evolutionary methods to docking of small molecules to protein targets,13 conformational analysis,14 drug dosing strategies,15 pharmacophore identification, similarity searching, derivation of quantitative structure−activity models, feature selection,16 and combinatorial library design.17 However, direct structural mutation of populations of druglike molecules or more complex materials, to optimize specific useful properties, is still in its relative infancy. Similarly, evolutionary methods, such as particle swarm optimization, have been applied successfully to de novo crystal structure prediction in various materials.18 Once structures have been derived, conventional electronic structure calculations can predict useful materials properties such as superconductivity or hardness.19 The application of evolutionary methods to these areas is not within the scope of the current review. Interested readers can consult the listed publications for more information on these applications of evolutionary methods. We have also not reviewed the application of evolutionary algorithms to the design of drugs, as this topic was covered recently in a mini-review.20 This review describes the most commonly used evolutionary methods for materials discovery and optimization and summarizes published examples of their application to identification of new and useful materials, or optimization of properties of existing materials. Using evolutionary approaches, the optimal conditions for chemical and materials synthesis, as well as the optimal materials properties, can be found quickly

2. EVOLUTIONARY ALGORITHMS Evolutionary algorithms (EAs) are generic, population-based, metaheuristic optimization methods. Metaheuristics are higherlevel procedures that generate approximate solutions to problems in parameter spaces, such as materials spaces, that are too large to be exhaustively sampled. The mechanisms by which EAs operate are inspired by biological evolutionary operations such as selection, mutation, recombination, and reproduction. Candidate solutions to the optimization problem play the role of population members, and a fitness function ranks the quality of the solutions. The evolutionary operators are applied to the top ranked members of the population, selection pressure is applied by a fitness function, and the evolutionary cycle is repeated until specified performance criteria are satisfied. Evolutionary algorithms are generic, as they do not make any assumptions about the nature of the fitness landscape. They have been applied widely across many areas of science and technology21 and in some engineering aspects of materials science22 but, paradoxically, not in materials design and optimization to a large extent. The main types of evolutionary algorithms are genetic algorithms (GAs), particle swarm optimization,23 genetic swarm optimization,23 ant colony optimization,24 evolutionary programming,25 and evolutionary strategies (more suitable for real value data). Whitley published an overview of evolutionary algorithms that describes practical issues and common pitfalls.26 Genetic algorithms dominate the applications of evolutionary algorithms in materials science, so we provide a more detailed summary of the method. 6108

DOI: 10.1021/acs.chemrev.5b00691 Chem. Rev. 2016, 116, 6107−6132

Chemical Reviews

Review

Figure 1. Steps in an evolutionary process.

2.1. Genetic Algorithms

The population size depends on the nature of the problem and, in the studies reviewed, varied between tens and several hundreds of materials. The way in which materials are presented in the genome can have a significant effect on the performance of the evolutionary algorithm, as was illustrated recently by Gobin and Schuth.28 An example of a materials genome used to optimize a property using genetic algorithms is given in Figure 2.

2.1.1. Overview of the Process. Genetic algorithms (GAs) were proposed by Holland9 and Rechenberg27 to solve combinatorial problems. As members of the broader class of evolutionary algorithms, they use techniques inspired by natural evolution, such as inheritance, mutation, selection, and crossover, to solve optimization problems. In the context of materials science, a genetic algorithm evolves a population of materials toward improved solutions to the optimization problem under study. Each material is represented mathematically by a set of properties (its chromosomes or genotype) that can be mutated and modified using specific operators. The initial population of individuals (materials) is often generated randomly but can also incorporate prior knowledge about suitable materials. This and subsequent populations (generations) are mutated to generate the next population in an iterative manner. The fitness of every individual in a given population is evaluated by means of a fitness or objective function, usually a desired and measurable property of the materials, but may also include undesirable properties to be avoided. The “fittest” materials are selected from the population, and their genomes are subjected to mutation operations to generate the new generation. A stochastic process is used to apply the mutational operations to selected population members. This iterative cycle continues until a maximum number of generations have been produced, or until some members of a population have properties that have exceeded an acceptance threshold. The basic steps in an evolutionary approach are summarized in Figure 1. The variables in a genetic optimization process include the materials genomes, the population sizes, the mutation operators used, the number of generations in the optimization, and the nature of the fitness function. 2.1.2. Materials Genomes. The evolutionary procedure starts with a population of different entities such as molecules or materials that are represented by genomes that encode compositional or structural parameters, but may also include information on how the material was manufactured or processed (provenance). Examples of elements of genomes include the mole fraction of individual components of the materials, the processing (e.g., calcination) temperature, presence or absence of specific promoter materials or acidity enhancers, polymer block sizes, monomer compositions, degree of cross-linking, and polydispersities. The elements of the genome can be continuous or integer or binary numbers. Most genetic representations of materials have a fixed length, although variable length representations are not precluded.

Figure 2. Example of a materials genome. Each of four metals used in a catalyst is represented by a binary string of seven bits. Converting the binary to decimal, the percentage contribution to the mole fraction can be calculated, and the material composition can be represented. Adapted from Umegaki et al.29 Copyright 2003 American Chemical Society.

2.1.3. Fitness Functions. The objective of any evolutionary optimization is to generate successive, relatively small populations of materials exhibiting substantial improvements in one or more specific, useful properties (the fitness or objective function(s)). The fitness function used to select the best performing population members is problem dependent and, for materials, is usually a property like catalyst efficiency, phosphor brightness or color, bioactivity, hardness, etc. As the evolutionary process proceeds, only a fraction of a given population is selected to breed the next generation. Selections are made by assessing the fitness of individuals, and a stochastic process (often a roulette wheel) is used to choose a mix of population members for the next generation, with fitter individuals being selected more frequently. The average fitness of each population generally increases because most of the best individuals from the preceding generation are selected for breeding. In some cases, multiple rather than single fitness functions are optimized. This usually leads to a compromise between the optimization objectives. In this scenario, there may exist a large number of nondominated or Pareto optimal solutions (those in which it is impossible to make any 6109

DOI: 10.1021/acs.chemrev.5b00691 Chem. Rev. 2016, 116, 6107−6132

Chemical Reviews

Review

or materials where it is possible to specify an appropriate genome that permits free movement on the fitness landscape.32 2.1.5. Genetic Operators. Genetic algorithms involve a set of operations: replication, mutation, and crossover. The mutation operations are applied to “parent” materials (individuals) selected for breeding from the pool. These genetic operators generate “child” materials that share characteristics of the “parents” but that are distinctly different. The replication (elitism) operator copies a genome into the next generation unchanged, a mutation operator sets one gene in the genome to a different value to obtain a new genome, and a crossover operator takes two or more genomes, splits them at some randomly defined point and then exchanges the pieces to build new genomes by mixing them according to certain rules (Figure 5). As there is a direct and deterministic relationship between the materials genome and its composition, structure, and provenance (e.g., processing conditions), it is a simple matter to translate the fittest “genotype” to the equivalent materials “phenotype”. The selection and mutation operations continue until a new population of materials of appropriate size is generated. The best (fittest) members of the population are selected and either carried forward unchanged into the next generation (elitism) or modified in some major or minor way by mutating the genomes of the materials before including them in the next generation. This new population of molecules or materials is then assessed against the fitness criteria and the cycle repeats until some stopping criteria (e.g., the materials have the desired property) are met. Other genetic algorithm “tuning” parameters include the mutation probability and the crossover probability

component better without making at least one other component worse, see Figure 3).

Figure 3. Illustration of a Pareto optimal surface. Any of the red points represents optimal choices. Other subjective factors may be used to choose between the points on the Pareto surface. Creative Commonsby-SA license (https://en.wikipedia.org/wiki/Pareto_efficiency).

2.1.4. Fitness Landscapes. A fitness landscape is a function in multidimensional space that describes the relationship between the fitness function and the genome. It is essentially a picture of the genotype−phenotype relationship, or an evaluation of a fitness function for all candidate solutions, a type of structure−property relationship. It is the surface that the evolutionary algorithms such as GAs explore when they optimize a fitness function by mutating a materials genome. It is also the surface that is approximated by machine learning models of the materials composition−fitness function relationship discussed below. The properties of fitness landscapes30,31 are usually dependent on two assumptions: the property used as the fitness function is physically well-defined; the genomes permit adequate movement on the landscape. Upon satisfaction of both assumptions, mathematical analysis shows that the topology (i.e., number and location of maxima, minima, and saddle points) of any fitness landscape should be monotonic; i.e., the landscape contains extremes only at globally maximal and/or minimal property values (as well as saddles at intermediate property values). Figure 4 illustrates fitness landscapes with two variables where the landscape contains multiple local maxima with different property values and is monotonic. Thus, evolutionary algorithms are applicable to any well-defined chemical or physical property and to all molecules

2.2. Structure−Property Model Methods as in Silico Fitness Functions

In many real-world optimization problems, the evaluation of the fitness of the population, essential for efficient optimization, is often performed experimentally. The time and expense of carrying out these experiments can sometimes dominate the optimization cost. The optimization process can be made faster and more cost efficient if prior information gained during the optimization process is used. The most obvious way to use this prior information is by building a model of the fitness function to assist in the selection of candidate solutions for evaluation. Such models are also called surrogates, metamodels, or approximation models and can be generated by any of the quantitative structure−property relationship (QSPR) methods commonly used for materials.5 These include linear and nonlinear regression, artificial neural networks (ANNs), support vector machines (SVMs), and various flavors of decision tree methods. ANNs have been used almost exclusively as surrogate fitness functions in the materials evolution literature, so we will discuss this modeling method in more detail. Artificial neural networks are versatile machine learning methods that have been widely used to predict diverse materials properties.5 They are supervised methods that use a set of training data to model the relationship between the composition, structures and processing parameters of materials (genotype) and their properties (phenotype). They are most useful when a relatively large amount of information is available to train them, such as data from high throughput experiments or prior information from genetic optimization experiments. The most commonly used ANNs are the feed-forward networks, which usually consist of three processing layers.

Figure 4. Visualization of a two-dimensional (two genes in the genome) fitness landscape. Different paths that the population could follow while mutating and evolving on the fitness landscape are shown in color. Creative Commons-by-SA license (https://en.wikipedia.org/ wiki/Fitness_landscape). 6110

DOI: 10.1021/acs.chemrev.5b00691 Chem. Rev. 2016, 116, 6107−6132

Chemical Reviews

Review

Figure 5. Example of materials genome encoding (e.g., Pn are process variables, Cn are composition variables), and common genetic algorithm operators.

Figure 6. Three-layer artificial neural network with five input nodes, three hidden layers, and one output node. More than one hidden layer can be used (deep neural network) and multiple output nodes can be specified where there is more than one property being modeled. The lower figure shows the structure of each nodeinputs are summed and passed through a linear or nonlinear transfer function to the next layer. Creative Commons-by-SA license 3.0. Stack Exchange (www.stackexchange.com).

The structure of a typical artificial neural network is shown in Figure 6. The input layer receives input data such as the composition, structure, or processing conditions of the materials or the materials genome, one or more hidden layers

perform nonlinear computation, and one or more output layers generate the predicted value of the property or fitness function (phenotype). It is common for the networks to have only one hidden layer, and the architecture of these networks is often 6111

DOI: 10.1021/acs.chemrev.5b00691 Chem. Rev. 2016, 116, 6107−6132

Chemical Reviews

6112

MATLAB implementation is available at http://www.mathworks.com/matlabcentral/fileexchange/10429-nsga-ii--amulti-objective-optimization-algorithm; C version is available at http://www.iitk.ac.in/kangal/codes.shtml MATLAB implementation is available on request from http://faculty.sites.uci.edu/jasper/sample/

http://dev.heuristiclab.com HeuristicLab

AMALGAM

examples of a simple genetic evolution problem codes in many different programming languages fast, nondominated multiobjective genetic algorithm; the metacode for the algorithm is presented by Deb et al.39 efficient version of NSGA-II; the metacode and description by Vrugt and Robinson40 package of many popular evolutionary and machine learning codes RosettaCode

NSGA-II

genetic programming toolbox for MATLAB GPLab

https://rosettacode.org/wiki/Evolutionary_algorithm

http://www.wardsystems.com/genehunter.asp

Ward Systems Group public domain public domain public domain public domain GNU public domain

http://gplab.sourceforge.net/features.html

http://au.mathworks.com/help/gads/genetic-algorithm.html Mathworks

global optimization toolbox; contains useful functions for developing MATLAB code for evolutionary optimization problems Microsoft Excel add-in that performs genetic optimization

https://sourceforge.net/directory/development/algorithms/genetic-algorithms/ public domain comprehensive list of largely public domain genetic algorithm software libraries, frameworks, and packages

SourceForge genetic algorithms MATLAB neural network GeneHunter

supplier

As the application of evolutionary algorithms such as genetic algorithms is at an early stage, almost all of the literature examples discussed below have used bespoke software to conduct the evolutional process. There are, however, a number of quite useful frameworks or packages that researchers interested in materials evolution might employ. Table 1 summarizes some of the more accessible evolutionary algorithm packages, many of which are in the public domain. A few of the literature studies that have not hand-coded their algorithms used the NSGA algorithm. Choice of parameters for genetic algorithm optimizations is problem dependent and is influenced by a number of factors. The optimum size of the population depends on factors such as the scale of experimentation, cost of materials and experiments, time and other resources, etc., and is a critical factor. Populations that are too large lead to a low convergence, while those that are too small are prone to converge toward a local optimum.35 The choice of mutation operators can also affect the search spaces. Crossover alters the materials genomes more drastically and can move searches into new areas of materials space, while point mutations tend to explore more local regions. The degree of discretization of the genomes is also an important parameter that is clearly also related to population sizes. For example, if the genome of a catalyst or phosphor is described by the mole fractions of the components and the processing temperature of the calcination, small step sizes in these parameters imply small search spaces (the ranges over which the parameters are varied) if the populations are kept small, and large populations if the search space is expanded. A useful strategy that is discussed at length below is the complementary use of quantitative structure−property

description

2.3. Genetic Algorithm Software and Optimization Parameter Choices

software

Table 1. Genetic Algorithm Computational Frameworks or Packages That Could Be Used in Materials Evolution Projects

described by the notation Nin − Nhidden − Nout, where Nin, Nhidden, and Nout are the numbers of nodes in the input, hidden, and output nodes, respectively. Deep neural networks with more than one hidden layer have also appeared recently, and they may offer significant advantages over traditional backpropagation neural networks.33,34 In all cases, care must be taken to avoid overtraining or overfitting problems. Nodes in the successive layers of the ANNs are connected to each other by weights. The inputs to the hidden and output nodes consist of the sum of the values of each input variable multiplied by the weight of all connections leading into the node. This sum is presented to a transfer function (linear for input and output nodes and sigmoidal for hidden nodes) to generate the output for that node. The difference between the measured value of the materials property and the network output is fed backward through the network to adjust the weights and minimize the error in the output. All materials in the data set are presented repeatedly to the network until the prediction performance is acceptable. Using an initial set of experimental data, ANNs can be trained and used to predict properties for all possible combinations so that the material with the optimum property value can be found, within the domain of applicability (DOA) of the models. The DOA is the parameter and property space in which the training data resides and in which the model makes the most reliable predictions. ANN models can therefore be very useful surrogate fitness functions that can be used to select the best members of populations of materials for mutation and progressing to the next evolutionary cycle.

Web address

Review

DOI: 10.1021/acs.chemrev.5b00691 Chem. Rev. 2016, 116, 6107−6132

Chemical Reviews

Review

relationship models embodied, for example, in a neural network. The model is trained on data generated over a useful range of materials space using a coarse discretization. The model can then interpolate in this region of space, effectively allowing a fine level of discretization to be used for subsequent evolutionary steps. As we discussed above, the materials search spaces are typically extremely large and not accessible by “brute force” combinatorial approaches. Although evolutionary methods such as genetic algorithms are among the most efficient methods to explore these potentially vast spaces, they clearly cannot explore all of it. Solutions will most likely be high fitness local optima on very complex response surfaces and are unlikely to be global optima. Nonetheless, like natural selection in biology, almost all of the local optima can be very useful and constitute substantial improvements over previous solutions. In addition, the size of search spaces and the potential complexity of response surfaces means that we cannot be sure that there is not a better local optimum relatively close by, unless we explore more widely. It is also important to consider constraints on evolution. Compositions clearly cannot have mole fractions that do not add up to 1, there are usually upper and lower bounds on processing temperatures or other parameters, abrupt phase or compositional changes may occur in some regions of composition space, and some materials may not be able to be synthesized at all. Some simple rules for synthetic accessibility have been proposed for small organic molecules,36 organometallics,37 etc. that could be used as constraints. In the case of phosphors discussed below,38 one of the constraints or fitness objectives used in the optimization was proximity to known phosphor compositions and the patent space embodied by these. Use of these factors allows the novelty of populations to be improved.

Figure 7. Selectivity of propene as a function of propane conversion for the first generation of catalytic materials (T = 500 °C, C3H8/O2/ N2 = 3/1/6). OF, objective function as criterion of catalyst quality to be minimized in the next generations. Reproduced with permission from Wolf et al.41 Copyright 2000 Elsevier.

The performance of all catalysts is summarized in Figure 8. The mean value of the objective function decreased (improved) from 107.7 to 98.3, and the propene yield increased from the first to the fourth generation. These authors reported an additional study on oxidative dehydrogenation of propane using catalysts containing other elements.42 The oxides they employed were V2O5, MoO3, MnO2, Fe2O3, ZnO, Ga2O3, GeO2, Nb2O5, WO3, Co3O4, CdO, In2O3, and NiO. The first generation consisted of 20 catalytic materials, each of which was composed of three elements. In the subsequent three generations, only 10 catalytic materials, seven of which were generated by mutation and three by crossover, were prepared and tested. The performance of the best four catalytic materials in each generation is illustrated in Figure 9, clearly illustrating the substantial increase in propene yield as the evolutionary procedure progressed. A similar study on optimizing catalysts for the oxidative dehydrogenation of propane was reported by Buyevskaya et al.43 Eight metal oxides, V2O5, MoO3, MnO2, Fe2O3, Ga2O3, La2O3, B2O3, and MgO, were used as primary components, and the genome consisted of a vector of the mole fractions used in the catalyst. Materials in the first generation consisted of four individual metal oxides deposited on α-alumina. The subsequent four generations of catalytic materials were created using mutation and crossover operators after assessing the fitness (propene yield) of each population member. Each of the generations consisted of 56 catalyst materials with different compositions. Figure 10 shows that the average propene yields increased from 3 to 6%, and the propene yields from the best catalysts increased from 7 to 9%, from the first to fifth generation, respectively. Evolutionary methods were also used by Rodemerck et al. to develop low-temperature catalysts for the total combustion of low-concentration propane in air.44 Such processes are essential for the fractional distillation of air to ensure that dangerous mixtures of liquid propane and liquid oxygen do not occur. The first generations of catalysts were created from two 60-member libraries of TiO2 supported and α-Fe2O3 supported catalysts generated by randomly mixing solutions of eight compounds (H2[PtCl6]·xH2O, (NH4)2PdCl6, RhCl3·2H2O, RuCl3·H2O, H[AuCl4]·3H2O, Ag lactate, Cu(NO3)2, Mn(NO3)2). Two subsequent generations, with population sizes of 45, were generated by applying mutation and crossover operators to the

3. EXAMPLES OF MATERIALS DISCOVERY AND OPTIMIZATION USING EVOLUTIONARY ALGORITHMS 3.1. Catalytic Materials

3.1.1. Catalyst Evolution Using Experimental Fitness Functions. The research area that has most frequently applied evolutionary approaches is that of catalyst identification and optimization. A relatively large number of research groups have reported using genetic algorithms to optimize the performance of these materials. The seminal paper on using the evolutionary approaches to optimize heterogeneous catalysts was that of Wolf et al.41 They aimed to optimize the conversion of propane to propene. Eight oxides (V2O5, MoO3, Fe2O3, GaO, MgO, B2O3, and La2O3) were chosen as primary catalyst components which, when randomly mixed and calcined, provided the first of four generations containing 56 catalysts each. Experiments showed that materials composed of four elements gave higher catalytic performance than materials with a greater or fewer number of elements. The objective (fitness) function used was the distance between a given catalyst performance (propene generated relative to propane consumed) to a hypothetical perfect catalyst (100% conversion of propane to propene), The selectivity of propene plotted against the propane conversion for all tested catalytic materials is given in Figure 7, together with an illustration of the objective function. Subsequent generations were created using mutation and crossover operators on the catalyst genomes, with vectors specifying the relative fraction of each element in the catalyst. 6113

DOI: 10.1021/acs.chemrev.5b00691 Chem. Rev. 2016, 116, 6107−6132

Chemical Reviews

Review

Figure 8. Left: Selectivity of propene as a function of propane conversion for all tested materials (T = 500 °C, C3H8/O2/N2 = 3/1/6). Mean OF: 107.7 (first generation), 105.5 (second generation), 100.3 (third generation), and 98.3 (fourth generation). Right: Propene yields of the 10 best catalysts in each generation. Reproduced with permission from Wolf et al.41 Copyright 2000 Elsevier.

compared to catalysts containing little or no Ru (30% CO yield were selected and their performances experimentally validated. The catalyst life was also optimized and ANNs were retrained using an additional three data points. The optimum catalyst suggested by the ANN model with fitted parameters a, b, and c as inputs had a much longer lifespan than the one found previously. A study aimed at developing catalysts for dry reforming of methane under pressure was also reported by the same research group. They used catalysts composed of cobalt and strontium carbonate with 10 different additives (B, K, Sc, Mn, Zn, Nb, Ag, Nd, Re, and Tl).85 The catalytic activities of 10 mol % Co + 1 mol % X/SrCO3 (X is one of the 10 additive elements) and 16 physicochemical properties of the 10 elements comprised the training data used to generate a neural network model for CO yield. The models predicted CO yield after 15 min or after 4 h 6128

DOI: 10.1021/acs.chemrev.5b00691 Chem. Rev. 2016, 116, 6107−6132

Chemical Reviews

Review

Figure 40. Results of the grid search for (a) CO conversion as functions of Co loading and pretreatment temperature (× is the maximum conversion) and (b) O2 conversion by H2 oxidation as a function of CO conversion. Adapted with permission from Omata et al.86 Copyright 2005 Elsevier.

selectivity for preferential oxidation at 200 °C and over 99% conversion of CO with the assistance of methanation of CO at 220 °C (Figure 40). The development of Co/SrCO 3 based catalysts for preferential oxidation of CO in excess H2 was also reported by Kobayashi et al.87 The catalysts consisted of 17 mol % Co + 1.7 mol % X/SrCO3 catalyst, where X was one of 10 elements (B, K, Sc, Mn, Zn, Nb, Ag, Nd, Re, and Tl). Sixteen physicochemical properties for 10 elements and the catalytic performance (CO conversion and O2 conversion by H2) were used to train a neural network model. The model was then used to predict the performance for 53 other additive elements. The effect of different additives was also analyzed using a multivariate data mining technique. The study identified a catalyst consisting of 17 mol % Co and 5.1 mol % B/SrCO3 catalyst that provided 99% CO conversion and 52% selectivity at 200 °C. Fine grained virtual screening of membrane disk catalysts for oxidative reforming of methane has been reported by Omata et al.88 They employed two porous alumina membrane filter disks in their catalysis rig, with one catalyst disk for methane combustion and one for methane reforming. The disks contained additives selected from nine elements (B, Ca, P, Mn, Fe, Cd, Ce, Gd, and Re). Experimental catalyst performance and 15 physicochemical properties for the additives were used to train the neural network models mapping catalyst composition to performance. The neural networks predicted the CH4 conversion and H2 selectivity. For the first disk, the model was used to predict the catalytic performance of NiZ, where Z was one additive element not used in training the model. Experiments confirmed that the predicted conversion and selectivity were close to those predicted by the neural network model. Among the predicted combinations, NiSc, NiY, and NiTb showed the highest conversion and selectivity. Similarly, for the second NiPr + CoMgZ based disks, the neural network was trained to predict their catalytic performance. An optimum composite for this disk consisted of NiPr plus CoMgLi that produced a CH4 conversion of 67%, an H2 selectivity of 82%, and a CO selectivity of 80%, almost the same level as two sheets of Rh alumina disk catalyst. Neural network models of objective functions and the NSGA-II genetic algorithm were used by Scott, Manos, and

Coveney to design electroceramic materials.89 They trained the neural network using literature data and that from a ceramic materials database. The structure−property models of permittivity, proximity to known ceramic compositions, and stoichiometry were used as multiobjective functions in a process where an evolutionary algorithm was used to “invert” the model to identify Pareto fronts for electroceramic materials with favorable properties. Very recently, a different quasi-evolutionary approach was reported by Thornton et al. They explored the performance limits for hydrogen storage for a very large virtual library of porous materials such as metal−organic frameworks, zeolitic imidazolate frameworks, and covalent organic frameworks.90 They employed computationally expensive grand canonical Monte Carlo methods using appropriate force fields to generate the structures and performances of a subset of the 850 000 possible structures, and then used neural networks to model the relationships between structural descriptors and performance. The neural network objective function models were then used to predict the evolutionary landscape and the fittest points in structure space used to generate new populations of improved porous materials. This process was iterated over three generations to identify porous materials with optimal hydrogen storage properties.

5. CONCLUSIONS AND PERSPECTIVE Evolutionary algorithms have found relatively wide applications in areas other than chemistry and materials. The increasing application of robotic and other high throughput methods for synthesis and characterization of materials provides a strongly enabling capability that will allow these very efficient methods to be applied to the discovery of new materials. These optimization methods are applicable to almost any type of materials that can be encoded into a genome, and they represent a fundamentally different approach to the general problem of materials design, identification, and optimization. Evolutionary algorithms should find increasing application in both the discovery and optimization of novel materials structures, and the optimization of synthesis and processing conditions that often play an important role in the properties of the final materials. The synergistic coupling of these powerful methods of exploring very large and multidimensional spaces with empirical machine learning models of fitness landscapes in 6129

DOI: 10.1021/acs.chemrev.5b00691 Chem. Rev. 2016, 116, 6107−6132

Chemical Reviews

Review

REFERENCES

the short to medium term, or high throughput ab initio calculations that predict materials properties in the long term, will allow materials scientists to find regions of very useful property space more quickly and effectively than by using high throughput synthesis and testing of materials alone. As Bohr (among others) once stated, it is difficult to make predictions, especially about the future. It is conceivable that evolutionary methods will have an impact on the large amount of research on self-assembling systems, resulting in methods to evolve self-assembling “smart” materials with unprecedented capabilities as delivery or targeting vehicles, sensors, etc. A glimpse of this future was provided by the work of Schulman,91 who used evolutionary methods to evolve the self-assembly of nano tiles, made from deterministically assembled DNA origami methods, into specific shapes. The Cronin group is also taking evolutionary methods in materials to a new level in their work on primordial chemistry, self-assembly, and evolution of oil droplets as model biochemical machines or to probe the pathways from complex chemical mixtures to autonomously evolving systems.92 The direction of this materials evolution research is taking generates a nice intellectual symmetry, given that powerful evolutionary search methods that mimic natural selection are being applied to materials that make up the biological genome (DNA) and the study of processes at the core of the development of life itself.

(1) Potyrailo, R.; Rajan, K.; Stoewe, K.; Takeuchi, I.; Chisholm, B.; Lam, H. Combinatorial and High-Throughput Screening of Materials Libraries: Review of State of the Art. ACS Comb. Sci. 2011, 13, 579− 633. (2) Xiang, X. D. Combinatorial Synthesis and High Throughput Evaluation of Functional Oxides - An Integrated Materials Chip Approach. Mater. Sci. Eng., B 1998, 56, 246−250. (3) Cerqueira, T. F. T.; Sarmiento-Perez, R.; Amsler, M.; Nogueira, F.; Botti, S.; Marques, M. A. L. Materials Design On-the-fly. J. Chem. Theory Comput. 2015, 11, 3955−3960. (4) Curtarolo, S.; Hart, G. L. W.; Nardelli, M. B.; Mingo, N.; Sanvito, S.; Levy, O. The High-Throughput Highway to Computational Materials Design. Nat. Mater. 2013, 12, 191−201. (5) Le, T.; Epa, V. C.; Burden, F. R.; Winkler, D. A. Quantitative Structure-Property Relationship Modeling of Diverse Materials Properties. Chem. Rev. 2012, 112, 2889−2919. (6) Forrest, S. Genetic Algorithms - Principles of Natural-Selection Applied to Computation. Science 1993, 261, 872−878. (7) Goldberg, D. E. Zen and the Art of Genetic Algorithms. Proceedings of the 3rd International Conference on Genetic Algorithms; Morgan Kaufmann: San Francisco, CA, 1989; pp 80−85. (8) Goldberg, D. E. Genetic Algorithm in Search, Optimization, and Machine Learning; Addison-Wesley: Reading, MA, 1989. (9) Holland, J. H. Adaptation in Natural and Artificial Systems; The University of Michigan Press: Ann Arbor, MI, 1975. (10) Diller, D. J. The Synergy between Combinatorial Chemistry and High-Throughput Screening. Curr. Opin Drug Discovery 2008, 11, 346−355. (11) Kodadek, T. The Rise, Fall and Reinvention of Combinatorial Chemistry. Chem. Commun. 2011, 47, 9757−9763. (12) Li, J. Q.; Ballmer, S. G.; Gillis, E. P.; Fujii, S.; Schmidt, M. J.; Palazzolo, A. M. E.; Lehmann, J. W.; Morehouse, G. F.; Burke, M. D. Synthesis of Many Different Types of Organic Small Molecules using One Automated Process. Science 2015, 347, 1221−1226. (13) Jones, G.; Willett, P.; Glen, R. C.; Leach, A. R.; Taylor, R. Development and Validation of a Genetic Algorithm for Flexible Docking. J. Mol. Biol. 1997, 267, 727−748. (14) Pavlov, T.; Todorov, M.; Stoyanova, G.; Schmieder, P.; Aladjov, H.; Serafimova, R.; Mekenyan, O. Conformational Coverage By a Genetic Algorithm: Saturation of Conformational Space. J. Chem. Inf. Model. 2007, 47, 851−863. (15) Bewick, S.; Yang, R.; Zhang, M. In 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society; IEEE: 2009; Vol. 1−20. (16) Yan, A.; Wang, Z.; Cai, Z. Prediction of Human Intestinal Absorption by GA Feature Selection and Support Vector Machine Regression. Int. J. Mol. Sci. 2008, 9, 1961−1976. (17) Clark, D. E.; Westhead, D. R. Evolutionary Algorithms in Computer-aided Molecular Design. J. Comput.-Aided Mol. Des. 1996, 10, 337−358. (18) Wang, Y. C.; Ma, Y. M. Perspective: Crystal Structure Prediction at High Pressures. J. Chem. Phys. 2014, 140, 040901. (19) Li, Y. W.; Hao, J.; Liu, H. Y.; Li, Y. L.; Ma, Y. M. The Metallization and Superconductivity of Dense Hydrogen Sulfide. J. Chem. Phys. 2014, 140, 174712. (20) Le, T. C.; Winkler, D. A. A Bright Future for Evolutionary Methods in Drug Design. ChemMedChem 2015, 10, 1296−1300. (21) Paszkowicz, W. Genetic Algorithms, a Nature-inspired Tool: Survey of Applications in Materials Science and Related Fields. Mater. Manuf. Processes 2009, 24, 174−197. (22) Paszkowicz, W. Genetic Algorithms, a Nature-inspired Tool: Survey of Applications in Materials Science and Related Fields: Part II. Mater. Manuf. Processes 2013, 28, 708−725. (23) Kennedy, J.; Eberhart, R. Particle Swarm Optimization. Proc. 1995 IEEE Int. Conf. Neural Networks 1995, 1−6, 1942−1948. (24) Blum, C. Ant Colony Optimization: Introduction and Recent Trends. Phys. Life Rev. 2005, 2, 353−373.

AUTHOR INFORMATION Corresponding Author

*E-mail: [email protected]. Notes

The authors declare no competing financial interest. Biographies Tu C. Le is a research scientist at CSIRO Manufacturing, Australia. She is working with Prof. David A. Winkler on modeling diverse materials properties, such as aqueous solubility, nanophase behavior of drug delivery systems, and flash points using machine-learning-based quantitative structure−property relationship modeling techniques. She is also involved in simulating guest−host systems and bioactive peptides using molecular dynamics methods. She completed her Ph.D. on hyperbranched polymer melt simulations at Swinburne University of Technology, Australia, in 2010 under the supervision of Prof. Billy D. Todd (Swinburne University of Technology), Prof. Peter Daivis (RMIT University), and Dr. Alfred Uhlherr (CSIRO). David A. Winkler is a senior principal research scientist with CSIRO Manufacturing and an adjunct professor at Monash, Latrobe, and Flinders Universities. His research interests have involved computational molecular design and complex systems. His recent research has focused on the effects of polymers, bioglasses, small bioactive species, and nanomaterials on biological systems. He works in regenerative medicine and collaborates with international stem cell biologists and tissue engineers. He employs computational methods developed for small molecule research to probe and predict the interactions of materials with complex biological systems. He was awarded traveling fellowships to Kyoto and Oxford and a Newton Turner Fellowship in 2009, and a Royal Academy of Engineering fellowship in 2015. He is a past board chairman of the Royal Australian Chemical Institute, past president of the Asian Federation for Medicinal Chemistry, and president-elect of the Federation of Asian Chemical Societies, and represents Australia on the organizing committee for Pacifichem. He has published over 200 scientific papers and patents. 6130

DOI: 10.1021/acs.chemrev.5b00691 Chem. Rev. 2016, 116, 6107−6132

Chemical Reviews

Review

(25) Fogel, L. J. Intelligence Through Simulated Evolution: Forty Years of Evolutionary Programming; Wiley: 1999. (26) Whitley, D. An Overview of Evolutionary Algorithms: Practical Issues and Common Pitfalls. Inf. Software Technol. 2001, 43, 817−831. (27) Rechenberg, I. Evolutionsstrategie: Optimierung Technischer Systeme nach Prinzipien der Biologischen Evolution; FrommannHolzboog: 1973. (28) Gobin, O. C.; Schüth, F. On the Suitability of Different Representations of Solid Catalysts for Combinatorial Library Design by Genetic Algorithms. J. Comb. Chem. 2008, 10, 835−846. (29) Umegaki, T.; Watanabe, Y.; Nukui, N.; Omata, K.; Yamada, M. Optimization of Catalyst for Methanol Synthesis by a Combinatorial Approach using a Parallel Activity Test and Genetic Algorithm Assisted by a Neural Network. Energy Fuels 2003, 17, 850−856. (30) Moore, K. W.; Pechen, A.; Feng, X. J.; Dominy, J.; Beltrani, V.; Rabitz, H. Universal Characteristics of Chemical Synthesis and Property Optimization. Chem. Sci. 2011, 2, 417−424. (31) Moore, K. W.; Pechen, A.; Feng, X. J.; Dominy, J.; Beltrani, V. J.; Rabitz, H. Why is Chemical Synthesis and Property Optimization Easier than Expected? Phys. Chem. Chem. Phys. 2011, 13, 10048− 10070. (32) Moore Tibbetts, K. W.; Li, R.; Pelczer, I.; Rabitz, H. Discovering Predictive Rules of Chemistry from Property Landscapes. Chem. Phys. Lett. 2013, 572, 1−12. (33) Ma, J. S.; Sheridan, R. P.; Liaw, A.; Dahl, G. E.; Svetnik, V. Deep Neural Nets as a Method for Quantitative Structure-Activity Relationships. J. Chem. Inf. Model. 2015, 55, 263−274. (34) LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436−444. (35) Llamas-Galilea, J.; Gobin, O. C.; Schuth, F. Comparison of Single- and Multiobjective Design of Experiment in Combinatorial Chemistry for the Selective Dehydrogenation of Propane. J. Comb. Chem. 2009, 11, 907−913. (36) Boda, K.; Seidel, T.; Gasteiger, J. Structure and Reaction Based Evaluation of Synthetic Accessibility. J. Comput.-Aided Mol. Des. 2007, 21, 311−325. (37) Foscato, M.; Occhipinti, G.; Venkatraman, V.; Alsberg, B. K.; Jensen, V. R. Automated Design of Realistic Organometallic Molecules from Fragments. J. Chem. Inf. Model. 2014, 54, 767−780. (38) Park, W. B.; Shin, N.; Hong, K. P.; Pyo, M.; Sohn, K. S. A New Paradigm for Materials Discovery: Heuristics-Assisted Combinatorial Chemistry Involving Parameterization of Material Novelty. Adv. Funct. Mater. 2012, 22, 2258−2266. (39) Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A Fast and Elitist Multiobjective Genetic Algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182−197. (40) Vrugt, J. A.; Robinson, B. A. Improved Evolutionary Optimization from Genetically Adaptive Multimethod Search. Proc. Natl. Acad. Sci. U. S. A. 2007, 104, 708−711. (41) Wolf, D.; Buyevskaya, O. V.; Baerns, M. An Evolutionary Approach in the Combinatorial Selection and Optimization of Catalytic Materials. Appl. Catal., A 2000, 200, 63−77. (42) Buyevskaya, O. V.; Wolf, D.; Baerns, M. Ethylene and Propene by Oxidative Dehydrogenation of Ethane and Propane - ‘Performance of Rare-Earth Oxide-Based Catalysts and Development of Redox-Type Catalytic Materials by Combinatorial Methods’. Catal. Today 2000, 62, 91−99. (43) Buyevskaya, O. V.; Bruckner, A.; Kondratenko, E. V.; Wolf, D.; Baerns, M. Fundamental and Combinatorial Approaches in the Search for and Optimisation of Catalytic Materials for the Oxidative Dehydrogenation of Propane to Propene. Catal. Today 2001, 67, 369−378. (44) Rodemerck, U.; Wolf, D.; Buyevskaya, O. V.; Claus, P.; Senkan, S.; Baerns, M. High-Throughput Synthesis and Screening of Catalytic Materials - Case Study on the Search for a Low-Temperature Catalyst for the Oxidation of Low-Concentration Propane. Chem. Eng. J. 2001, 82, 3−11. (45) Yamada, Y.; Ueda, A.; Nakagawa, K.; Kobayashi, T. Optimization Of Fe/SiO2 Based Metal Oxides as Selective Oxidation

Catalyst of Propane with Combinatorial Approach. Res. Chem. Intermed. 2002, 28, 397−407. (46) Corma, A.; Serra, J. M.; Chica, A. Discovery of New Paraffin Isomerization Catalysts Based on SO42‑/ZrO2 And WOx/ZrO2 Applying Combinatorial Techniques. Catal. Today 2003, 81, 495−506. (47) Serra, J. M.; Chica, A.; Corma, A. Development of a Low Temperature Light Paraffin Isomerization Catalysts with Improved Resistance to Water and Sulphur by Combinatorial Methods. Appl. Catal., A 2003, 239, 35−42. (48) Kirsten, G.; Maier, W. F. Strategies for the Discovery of New Catalysts with Combinatorial Chemistry. Appl. Surf. Sci. 2004, 223, 87−101. (49) Paul, J. S.; Janssens, R.; Denayer, J. F. M.; Baron, G. V.; Jacobs, P. A. Optimization of MoVSb Oxide Catalyst for Partial Oxidation of Isobutane by Combinatorial Approaches. J. Comb. Chem. 2005, 7, 407−413. (50) Moehmel, S.; Steinfeldt, N.; Engelschalt, S.; Holena, M.; Kolf, S.; Baerns, A.; Dingerdissen, U.; Wolf, D.; Weber, R.; Bewersdorf, M. New Catalytic Materials for the High-Temperature Synthesis of Hydrocyanic Acid from Methane and Ammonia by High-Throughput Approach. Appl. Catal., A 2008, 334, 73−83. (51) Zitzler, E.; Laumanns, M.; Thiele, L. EUROGEN 2001; International Center for Numerical Methods in Engineering (CIMNE): 2011; pp 95−100. (52) Kreutz, J. E.; Shukhaev, A.; Du, W. B.; Druskin, S.; Daugulis, O.; Ismagilov, R. F. Evolution of Catalysts Directed by Genetic Algorithms in a Plug-Based Microfluidic Device Tested with Oxidation of Methane by Oxygen. J. Am. Chem. Soc. 2010, 132, 3128−3132. (53) Labinger, J. A.; Bercaw, J. E. Understanding and Exploiting C-H Bond Activation. Nature 2002, 417, 507−514. (54) Bergman, R. G.; Cundari, T. R.; Gillespie, A. M.; Gunnoe, T. B.; Harman, W. D.; Klinckman, T. R.; Temple, M. D.; White, D. P. Computational Study of Methane Activation By TpRe(Co)2 And CpRe(Co)2 with a Stereoelectronic Comparison of Cyclopentadienyl and Scorpionate Ligands. Organometallics 2003, 22, 2331−2337. (55) Thomas, S.; Dawe, R. A. Review of Ways to Transport Natural Gas Energy from Countries Which do not Need the Gas for Domestic Use. Energy 2003, 28, 1461−1477. (56) Bar-Nahum, I.; Khenkin, A. M.; Neumann, R. Mild, Aqueous, Aerobic, Catalytic Oxidation of Methane to Methanol and Acetaldehyde Catalyzed by a Supported BipyrimidinylplatinumPolyoxometalate Hybrid Compound. J. Am. Chem. Soc. 2004, 126, 10236−10237. (57) Rodemerck, U.; Holena, M.; Wagner, E.; Smejkal, Q.; Barkschat, A.; Baerns, M. Catalyst Development for CO2 Hydrogenation To Fuels. ChemCatChem 2013, 5, 1948−1955. (58) Watanabe, Y.; Umegaki, T.; Hashimoto, M.; Omata, K.; Yamada, M. Optimization of Cu Oxide Catalysts for Methanol Synthesis by Combinatorial Tools using 96 Well Microplates, Artificial Neural Network and Genetic Algorithm. Catal. Today 2004, 89, 455− 464. (59) Huang, K.; Zhan, X. L.; Chen, F. Q.; Lu, D. W. Catalyst Design for Methane Oxidative Coupling by Using Artificial Neural Network and Hybrid Genetic Algorithm. Chem. Eng. Sci. 2003, 58, 81−87. (60) Huang, K.; Chen, F. Q.; Lu, D. W. Artificial Neural NetworkAided Design of a Multi-Component Catalyst for Methane Oxidative Coupling. Appl. Catal., A 2001, 219, 61−68. (61) Corma, A.; Serra, J. M.; Serna, P.; Valero, S.; Argente, E.; Botti, V. Optimisation of Olefin Epoxidation Catalysts with the Application of High-Throughput and Genetic Algorithms Assisted by Artificial Neural Networks (Softcomputing Techniques). J. Catal. 2005, 229, 513−524. (62) Rodemerck, U.; Baerns, M.; Holena, M.; Wolf, D. Application of a Genetic Algorithm and a Neural Network for the Discovery and Optimization of New Solid Catalytic Materials. Appl. Surf. Sci. 2004, 223, 168−174. (63) Froemming, N. S.; Henkelman, G. Optimizing Core-Shell Nanoparticle Catalysts with a Genetic Algorithm. J. Chem. Phys. 2009, 131, 234103. 6131

DOI: 10.1021/acs.chemrev.5b00691 Chem. Rev. 2016, 116, 6107−6132

Chemical Reviews

Review

(64) Rabitz, H. Control in the Sciences over Vast Length and Time Scales. Quant. Phys. Lett. 2012, 1−19. (65) Sohn, K. S.; Kim, B. I.; Shin, N. Genetic Algorithm-Assisted Combinatorial Search for New Red Phosphors of High Efficiency at Soft Ultraviolet Excitation. J. Electrochem. Soc. 2004, 151, H243−H248. (66) Sohn, K. S.; Park, D. H.; Cho, S. H.; Kim, B. I.; Woo, S. I. Genetic Algorithm-Assisted Combinatorial Search for a New Green Phosphor for Use in Tricolor White LEDs. J. Comb. Chem. 2006, 8, 44−49. (67) Jung, Y. S.; Kulshreshtha, C.; Kim, J. S.; Shin, N.; Sohn, K. S. Genetic Algorithm-Assisted Coinabinatorial Search for New Blue Phosphors in a (Ca,Sr,Ba,Mg,Eu)xByPzOδ System. Chem. Mater. 2007, 19, 5309−5318. (68) Kulshreshtha, C.; Sharma, A. K.; Sohn, K. S. Search for New Red Phosphors using Genetic Algorithm-Assisted Combinatorial Chemistry. J. Comb. Chem. 2008, 10, 421−425. (69) Sharma, A. K.; Kulshreshtha, C.; Sohn, K. S. Discovery of New Green Phosphors and Minimization of Experimental Inconsistency using a Multi-Objective Genetic Algorithm-Assisted Combinatorial Method. Adv. Funct. Mater. 2009, 19, 1705−1712. (70) Hirosaki, N.; Xie, R. J.; Kimoto, K.; Sekiguchi, T.; Yamamoto, Y.; Suehiro, T.; Mitomo, M. Characterization and Properties of GreenEmitting β-SiAlON:Eu2+ Powder Phosphors for White Light-Emitting Diodes. Appl. Phys. Lett. 2005, 86, 211905. (71) Li, Y. Q.; Delsing, A. C. A.; de With, G.; Hintzen, H. T. Luminescence Properties of Eu2+-Activated Alkaline-Earth SiliconOxynitride MSi2O2‑δN2+2/3δ (M = Ca, Sr, Ba): A Promising Class of Novel LED Conversion Phosphors. Chem. Mater. 2005, 17, 3242− 3248. (72) Uheda, K.; Hirosaki, N.; Yamamoto, Y.; Naito, A.; Nakajima, T.; Yamamoto, H. Luminescence Properties of a Red Phosphor, CaAlSiN3: Eu2+, for White Light-Emitting Diodes. Electrochem. SolidState Lett. 2006, 9, H22−H25. (73) Xie, R. J.; Hirosaki, N.; Suehiro, T.; Xu, F. F.; Mitomo, M. A Simple, Efficient Synthetic Route To Sr2Si5N8: Eu2+-Based Red Phosphors for White Light-Emitting Diodes. Chem. Mater. 2006, 18, 5578−5583. (74) Omata, K.; Ozaki, T.; Umegaki, T.; Watanabe, Y.; Nukui, N.; Yamada, M. Optimization of the Temperature Profile of a Temperature Gradient Reactor for DME Synthesis using a Simple Genetic Algorithm Assisted by a Neural Network. Energy Fuels 2003, 17, 836− 841. (75) Bulut, M.; Gevers, L. E. M.; Paul, J. S.; Vankelecom, N. F. J.; Jacobs, P. A. Directed Development of High-Performance Membranes via High-Throughput and Combinatorial Strategies. J. Comb. Chem. 2006, 8, 168−173. (76) Yelk, J.; Sukharev, M.; Seideman, T. Optimal Design of Nanoplasmonic Materials using Genetic Algorithms as a Multiparameter Optimization Tool. J. Chem. Phys. 2008, 129, 064706. (77) Ginzburg, P.; Berkovitch, N.; Nevet, A.; Shor, I.; Orenstein, M. Resonances On-Demand for Plasmonic Nano-Particles. Nano Lett. 2011, 11, 2329−2333. (78) Kim, S.; Sohn, K. S.; Pyo, M. Genetic Algorithm-Assisted Optimization of Nanoporous TiO2 for Low-Temperature Processable Photoanodes of Dye-Sensitized Solar Cells. ACS Comb. Sci. 2011, 13, 101−106. (79) Bae, E. G.; Kim, H.; Hwang, Y. H.; Sohn, K. S.; Pyo, M. Genetic Algorithm-Assisted Optimization of Partially Dyed-TiO2 for RoomTemperature Printable Photoanodes of Dye-Sensitized Solar Cells. J. Mater. Chem. 2012, 22, 551−556. (80) Mohid, M.; Miller, J. F.; Harding, S. L.; Tufte, G.; Lykkebo, O. R.; Massey, M. K.; Petty, M. C. Evolution-In-Materio: Solving Function Optimization Problems using Materials. 14th UK Workshop Comput. Intell. 2014, 7−14. (81) Parrill, A. L. Evolutionary and Genetic Methods in Drug Design. Drug Discovery Today 1996, 1, 514−521. (82) Omata, K.; Hashimoto, M.; Watanabe, Y.; Umegaki, T.; Wagatsuma, S.; Ishiguro, G.; Yamada, M. Optimization of Cu Oxide

Catalyst for Methanol Synthesis under High CO2 Partial Pressure using Combinatorial Tools. Appl. Catal., A 2004, 262, 207−214. (83) Omata, K.; Nukui, N.; Yamada, M. Artificial Neural Network Aided Design of a Stable Co-MgO Catalyst of High-Pressure Dry Reforming of Methane. Ind. Eng. Chem. Res. 2005, 44, 296−301. (84) Hirose, K.; Ueda, T. Introduction To Taguchi Method Analysis; Doyukan: Tokyo, 2003. (85) Omata, K.; Kobayashi, Y.; Yamada, M. Artificial Neural Network Aided Screening and Optimization of Additives to Co/SrCO3 Catalyst for Dry Reforming of Methane under Pressure. Energy Fuels 2009, 23, 1931−1935. (86) Omata, K.; Kobayashi, Y.; Yamada, M. Artificial Neural Network-Aided Development of Supported Co Catalyst for Preferential Oxidation of CO in Excess Hydrogen. Catal. Commun. 2005, 6, 563−567. (87) Kobayashi, Y.; Omata, K.; Yamada, M. Screening of Additives to a Co/SrCO3 Catalyst by Artificial Neural Network for Preferential Oxidation of CO In Excess H2. Ind. Eng. Chem. Res. 2010, 49, 1541− 1549. (88) Omata, K.; Masuda, A.; Ishii, H.; Suzuki, H.; Yamada, M. Artificial Neural Network Aided Screening for Membrane Disc Catalysts for Oxidative Reforming of Methane. Appl. Catal., A 2009, 362, 14−19. (89) Scott, D. J.; Manos, S.; Coveney, P. V. Design of Electroceramic Materials using Artificial Neural Networks and Multiobjective Evolutionary Algorithms. J. Chem. Inf. Model. 2008, 48, 262−273. (90) Thornton, A. W.; Simon, C.; Kim, J. S.; Deeg, K. S.; Konstas, K.; Hill, M. R.; Winkler, D. A.; Haranczyk, M.; Smit, B. The Materials Genome in Action: Identifying the Performance Limits of Hydrogen Storage. Energy Environ. Sci. 2016. (91) Schulman, R. Beyond biology: Designing a New Mechanism for Self-Replication and Evolution at the Nanoscale. Proc. 13th Annu. Genet. Evol. Comput. Conf. 2011, 7−14. (92) Gutierrez, J. M. P.; Hinkley, T.; Taylor, J. W.; Yanev, K.; Cronin, L. Evolution of Oil Droplets in a Chemorobotic Platform. Nat. Commun. 2014, 5, 5571.

6132

DOI: 10.1021/acs.chemrev.5b00691 Chem. Rev. 2016, 116, 6107−6132