使用Python优化神经网络参数的遗传算法
人工神经网络是一种监督机器学习算法,在语音和图像识别、时间序列预测、机器翻译软件等领域都有广泛的应用。它们在研究中很有用,因为它们能够解决随机问题,而随机问题通常允许对极其复杂的问题求近似解。
然而,很难定义理想的网络结构,因为中间层中有多少神经元,中间层中有多少神经元,以及这些神经元之间的连接应该如何实现都没有明确的规则。为了解决这类问题,本文将指导如何使用遗传算法在Python中自动查找良好的神经网络体系结构。
首先,您需要安装scikit-learn软件包。
为了训练混合算法,我们将使用Iris机器学习数据集。
from sklearn import datasets import numpy as np import matplotlib.pyplot as plt from sklearn.neural_network import MLPClassifier from sklearn.metrics import accuracy_score from sklearn.model_selection import train_test_split from random import randint import random from sklearn.metrics import mean_absolute_error as mae iris = datasets.load_iris() X = iris.data y = iris.target X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.33, random_state=42)
现在我们可以开始构建遗传算法。这里的神经网络有两个隐藏层。下面的Python代码显示了种群初始化的示例。种群大小由size_mlp定义。
def inicialization_populacao_mlp(size_mlp): pop = [[]]*size_mlp activation = ['identity','logistic', 'tanh', 'relu'] solver = ['lbfgs','sgd', 'adam'] pop = [[random.choice(activation), random.choice(solver), randint(2,100), randint(2,100)] for i in range(0, size_mlp)] return pop
交叉算子是一种将双亲的信息结合起来产生新的个体的算子。目标是增加遗传变异,并提供更好的选择。这里使用单点交叉。
def crossover_mlp(mother_1, mother_2): child = [mother_1[0], mother_2[1], mother_1[2], mother_2[3]] return child
为了进一步增加遗传变异和避免局部极小值,使用的另一个算子是突变。突变的概率由prob_mut定义。
def mutation_mlp(child, prob_mut): for c in range(0, len(child)): if np.random.rand() > prob_mut: k = randint(2,3) child[c][k] = int(child[c][k]) + randint(1, 10) return child
因为此示例是分类任务,所以适应度函数是根据神经网络的准确性计算的,在这种情况下,遗传算法的目标是使神经网络的准确性最大化。
def function_fitness_mlp(pop, X_train, y_train, X_test, y_test, size_mlp): fitness = [[]]*size_mlp classifiers = [[]]*size_mlp j = 0 for w in pop: clf = MLPClassifier(activation=w[0], solver=w[1], alpha=1e-5, hidden_layer_sizes=(int(w[2]), int(w[3])), random_state=1) clf.fit(X_train, y_train) fitness[j] = mae(clf.predict(X_test), y_test) classifiers[j] = clf j = j+1 return fitness, classifiers
最后,遗传算法的主体得以构建。
def ag_mlp(X_train, y_train, X_test, y_test, num_epochs = 10, size_mlp=10, prob_mut=0.5): pop = inicialization_populacao_mlp(size_mlp) fitness, classifiers = function_fitness_mlp(pop, X_train, y_train, X_test, y_test, size_mlp) pop_fitness = np.array(list(zip(pop, fitness, classifiers))) pop_fitness_sort = pop_fitness[pop_fitness[:,1].argsort()] # population initialization for j in range(0, num_epochs): #seleciona os pais parent_1 = pop_fitness_sort[0:int(size_mlp/2)][:,0] parent_2 = pop_fitness_sort[int(size_mlp/2)::][:,0] #cruzamento child_1 = [crossover_mlp(parent_1[i], parent_2[i]) for i in range(0, len(parent_1))] child_1 = np.array(list(map(list, child_1))) child_2 = [crossover_mlp(parent_2[i], parent_1[i]) for i in range(0, len(parent_1))] child_2 = np.array(list(map(list, child_2))) child_1 = mutation_mlp(child_1, prob_mut) child_2 = mutation_mlp(child_2, prob_mut) #calculates children's fitness to choose who will move on to the next generation fitness_child_1, classifiers_child_1 = function_fitness_mlp(child_1,X_train, y_train, X_test, y_test, size_mlp) fitness_child_2, classifiers_child_2 = function_fitness_mlp(child_2, X_train, y_train, X_test, y_test, size_mlp) fitness_child_1 = np.array(list(zip(child_1, fitness_child_1, classifiers_child_1))) fitness_child_2 = np.array(list(zip(child_2, fitness_child_2, classifiers_child_2))) #selects next generation individuals pop_all = np.concatenate((fitness_child_1, fitness_child_2, pop_fitness), axis=0) pop_all_sort = pop_all[pop_all[:,1].argsort()] best_individual = pop_all_sort[0] pop_fitness_sort = pop_all_sort[0:size_mlp] return pop_fitness_sort[0]
相关推荐
oXiaoChong 2020-10-26
taiyangshenniao 2020-06-11
巴壁虎 2020-05-17
风吹夏天 2020-05-18
rein0 2020-04-21
seekerhit 2020-01-29
Happyunlimited 2020-01-29
laohyx 2020-01-23
ustbfym 2019-12-30
ustbfym 2019-11-08
lixiaotao 2019-11-03
yishujixiaoxiao 2019-11-03
风吹夏天 2014-08-12
Broadview 2019-06-27
littletingting 2019-06-19
liuxiaohua 2019-06-20
duyifei0 2019-06-20
liuxiaohua 2019-06-20
hellosunshine 2017-11-14