麻豆小视频在线观看_中文黄色一级片_久久久成人精品_成片免费观看视频大全_午夜精品久久久久久久99热浪潮_成人一区二区三区四区

首頁 > 編程 > Python > 正文

使用TensorFlow實現SVM

2020-02-15 22:55:17
字體:
來源:轉載
供稿:網友

較基礎的SVM,后續會加上多分類以及高斯核,供大家參考。

Talk is cheap, show me the code

import tensorflow as tffrom sklearn.base import BaseEstimator, ClassifierMixinimport numpy as npclass TFSVM(BaseEstimator, ClassifierMixin): def __init__(self,   C = 1, kernel = 'linear',   learning_rate = 0.01,   training_epoch = 1000,   display_step = 50,  batch_size = 50,  random_state = 42):  #參數列表  self.svmC = C  self.kernel = kernel  self.learning_rate = learning_rate  self.training_epoch = training_epoch  self.display_step = display_step  self.random_state = random_state  self.batch_size = batch_size def reset_seed(self):  #重置隨機數  tf.set_random_seed(self.random_state)  np.random.seed(self.random_state) def random_batch(self, X, y):  #調用隨機子集,實現mini-batch gradient descent  indices = np.random.randint(1, X.shape[0], self.batch_size)  X_batch = X[indices]  y_batch = y[indices]  return X_batch, y_batch def _build_graph(self, X_train, y_train):  #創建計算圖  self.reset_seed()  n_instances, n_inputs = X_train.shape  X = tf.placeholder(tf.float32, [None, n_inputs], name = 'X')  y = tf.placeholder(tf.float32, [None, 1], name = 'y')  with tf.name_scope('trainable_variables'):   #決策邊界的兩個變量   W = tf.Variable(tf.truncated_normal(shape = [n_inputs, 1], stddev = 0.1), name = 'weights')   b = tf.Variable(tf.truncated_normal([1]), name = 'bias')  with tf.name_scope('training'):   #算法核心   y_raw = tf.add(tf.matmul(X, W), b)   l2_norm = tf.reduce_sum(tf.square(W))   hinge_loss = tf.reduce_mean(tf.maximum(tf.zeros(self.batch_size, 1), tf.subtract(1., tf.multiply(y_raw, y))))   svm_loss = tf.add(hinge_loss, tf.multiply(self.svmC, l2_norm))   training_op = tf.train.AdamOptimizer(learning_rate = self.learning_rate).minimize(svm_loss)  with tf.name_scope('eval'):   #正確率和預測   prediction_class = tf.sign(y_raw)   correct_prediction = tf.equal(y, prediction_class)   accuracy = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))  init = tf.global_variables_initializer()  self._X = X; self._y = y  self._loss = svm_loss; self._training_op = training_op  self._accuracy = accuracy; self.init = init  self._prediction_class = prediction_class  self._W = W; self._b = b def _get_model_params(self):  #獲取模型的參數,以便存儲  with self._graph.as_default():   gvars = tf.get_collection(tf.GraphKeys.GLOBAL_VARIABLES)  return {gvar.op.name: value for gvar, value in zip(gvars, self._session.run(gvars))} def _restore_model_params(self, model_params):  #保存模型的參數  gvar_names = list(model_params.keys())  assign_ops = {gvar_name: self._graph.get_operation_by_name(gvar_name + '/Assign') for gvar_name in gvar_names}  init_values = {gvar_name: assign_op.inputs[1] for gvar_name, assign_op in assign_ops.items()}  feed_dict = {init_values[gvar_name]: model_params[gvar_name] for gvar_name in gvar_names}  self._session.run(assign_ops, feed_dict = feed_dict) def fit(self, X, y, X_val = None, y_val = None):  #fit函數,注意要輸入驗證集  n_batches = X.shape[0] // self.batch_size  self._graph = tf.Graph()  with self._graph.as_default():   self._build_graph(X, y)  best_loss = np.infty  best_accuracy = 0  best_params = None  checks_without_progress = 0  max_checks_without_progress = 20  self._session = tf.Session(graph = self._graph)  with self._session.as_default() as sess:   self.init.run()   for epoch in range(self.training_epoch):    for batch_index in range(n_batches):     X_batch, y_batch = self.random_batch(X, y)     sess.run(self._training_op, feed_dict = {self._X:X_batch, self._y:y_batch})    loss_val, accuracy_val = sess.run([self._loss, self._accuracy], feed_dict = {self._X: X_val, self._y: y_val})    accuracy_train = self._accuracy.eval(feed_dict = {self._X: X_batch, self._y: y_batch})    if loss_val < best_loss:     best_loss = loss_val     best_params = self._get_model_params()     checks_without_progress = 0    else:     checks_without_progress += 1     if checks_without_progress > max_checks_without_progress:      break    if accuracy_val > best_accuracy:     best_accuracy = accuracy_val     #best_params = self._get_model_params()    if epoch % self.display_step == 0:     print('Epoch: {}/tValidaiton loss: {:.6f}/tValidation Accuracy: {:.4f}/tTraining Accuracy: {:.4f}'      .format(epoch, loss_val, accuracy_val, accuracy_train))   print('Best Accuracy: {:.4f}/tBest Loss: {:.6f}'.format(best_accuracy, best_loss))   if best_params:    self._restore_model_params(best_params)    self._intercept = best_params['trainable_variables/weights']    self._bias = best_params['trainable_variables/bias']   return self def predict(self, X):  with self._session.as_default() as sess:   return self._prediction_class.eval(feed_dict = {self._X: X}) def _intercept(self):  return self._intercept def _bias(self):  return self._bias            
發表評論 共有條評論
用戶名: 密碼:
驗證碼: 匿名發表
主站蜘蛛池模板: 国产精品久久久久久久久久东京 | 国产精品久久久久久久久久 | 极品大长腿啪啪高潮露脸 | 最新午夜综合福利视频 | 精品一区二区视频在线观看 | 国产女同疯狂激烈互摸 | 羞羞的| 狠狠干视频网站 | 久久久在线| 狠狠干夜夜草 | 中文字幕电影免费播放 | 国产精品久久久久久久久久东京 | 免费永久在线观看黄网 | 毛毛片在线看 | 水多视频在线观看 | 欧美日本日韩 | 欧美77| 国产精品视频一区二区三区四 | 欧美a∨一区二区三区久久黄 | 亚洲片在线观看 | 91精品国产乱码久久久久久久久 | 538任你躁在线精品视频网站 | 双性精h调教灌尿打屁股的文案 | 国产精品一区二区免费在线观看 | 在线播放黄色片 | 一级做人爱c黑人影片 | 久久久久久久久久亚洲精品 | 爱射av| 欧美精品一区自拍a毛片在线视频 | 调教小男生抽打尿孔嗯啊视频 | 密室逃脱第一季免费观看完整在线 | 人人舔人人插 | a一级黄色大片 | 国产精品久久久久久久久久久久久久久久 | 黄污污网站 | 性生活视频软件 | 国产免费久久久久 | 妇子乱av一区二区三区 | 色污视频| 亚洲白嫩在线观看 | 一级@片 |