sklearn系列学习--支持向量机SVM

版权声明:本文为博主原创文章,未经博主允许不得转载。 https://blog.csdn.net/yangfengling1023/article/details/83000899
#coding:utf-8
##1 导入svm和数据集
from sklearn import svm,datasets

##2 调用SVC()
clf = svm.SVC()

##3 载入鸢尾花数据集
iris = datasets.load_iris()
X = iris.data
y = iris.target

##4 fit()训练
clf.fit(X,y)

##5 predict()预测
pre_y = clf.predict(X[5:10])
print(pre_y)
print(y[5:10])

##6 导入numpy
import numpy as np
test = np.array([[5.1,2.9,1.8,3.6]])

##7 对test进行预测
test_y = clf.predict(test)
print(test_y)

运行结果如下所示:

本博客内容会继续更新,欢迎交流

keras还没有官方实现attention机制,有些attention的个人实现,在mnist数据集上做了下实验。模型是双向lstm+attention+dropout,话说双向lstm本身就很强大了。 
参考链接:https://github.com/philipperemy/keras-attention-mechanism 
https://github.com/keras-team/keras/issues/1472 

代码如下:
import keras
import numpy as np
np.random.seed(1337)
from keras.datasets import mnist
from keras.utils import np_utils
from keras.layers import *
from keras.models import *
from keras.optimizers import Adam

TIME_STEPS = 28
INPUT_DIM = 28
lstm_units = 64

# data pre-processing
(X_train, y_train), (X_test, y_test) = mnist.load_data('mnist.npz')
X_train = X_train.reshape(-1, 28, 28) / 255.
X_test = X_test.reshape(-1, 28, 28) / 255.
y_train = np_utils.to_categorical(y_train, num_classes=10)
y_test = np_utils.to_categorical(y_test, num_classes=10)
print('X_train shape:', X_train.shape)
print('X_test shape:', X_test.shape)

# first way attention
def attention_3d_block(inputs):
    #input_dim = int(inputs.shape[2])
    a = Permute((2, 1))(inputs)
    a = Dense(TIME_STEPS, activation='softmax')(a)
    a_probs = Permute((2, 1), name='attention_vec')(a)
    #output_attention_mul = merge([inputs, a_probs], name=’attention_mul’, mode=’mul’)
    output_attention_mul = multiply([inputs, a_probs], name='attention_mul')
    return output_attention_mul

# build RNN model with attention
inputs = Input(shape=(TIME_STEPS, INPUT_DIM))
drop1 = Dropout(0.3)(inputs)
lstm_out = Bidirectional(LSTM(lstm_units, return_sequences=True), name='bilstm')(drop1)
attention_mul = attention_3d_block(lstm_out)
attention_flatten = Flatten()(attention_mul)
drop2 = Dropout(0.3)(attention_flatten)
output = Dense(10, activation='sigmoid')(drop2)
model = Model(inputs=inputs, outputs=output)

# second way attention
# inputs = Input(shape=(TIME_STEPS, INPUT_DIM))
# units = 32
# activations = LSTM(units, return_sequences=True, name=’lstm_layer’)(inputs)
#
# attention = Dense(1, activation=’tanh’)(activations)
# attention = Flatten()(attention)
# attention = Activation(‘softmax’)(attention)
# attention = RepeatVector(units)(attention)
# attention = Permute([2, 1], name=’attention_vec’)(attention)
# attention_mul = merge([activations, attention], mode=’mul’, name=’attention_mul’)
# out_attention_mul = Flatten()(attention_mul)
# output = Dense(10, activation=’sigmoid’)(out_attention_mul)
# model = Model(inputs=inputs, outputs=output)

model.compile(optimizer='adam', loss='categorical_crossentropy', metrics=['accuracy'])
print(model.summary())

print('Training————')
model.fit(X_train, y_train, epochs=10, batch_size=16)

print('Testing————–')
loss, accuracy = model.evaluate(X_test, y_test)

print('test loss:', loss)
print('test accuracy:', accuracy)

猜你喜欢

转载自blog.csdn.net/yangfengling1023/article/details/83000899